bigscience-workshop/petals
🌸 Run LLMs at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading
PythonMIT
Pinned issues
Issues
- 5
- 0
- 3
- 1
Add YaLM-100B
#512 opened - 4
[might be bug?] Failed to connect to bootstrap peers when using docker image on truenas scale
#511 opened - 1
- 1
- 1
- 2
Issue with beam search decoding
#503 opened - 1
More powerful session API
#495 opened - 1
Incentive system based on Lightning
#494 opened - 0
- 9
Can't install on windows
#488 opened - 2
How can we make this work long term?
#483 opened - 1
`model.generate(input_ids=...)` support
#481 opened - 2
Python the only option to run?
#469 opened - 0
- 1
how to avoid this server failure? Seems to happen randomly after 1 hour of running a script.
#466 opened - 4
Out of memory on a client with 8 GB RAM
#465 opened - 3
- 0
Add server option to prioritize a list of given peer IDs (so you can prioritize your own clients)
#457 opened - 0
Hope to get in touch
#455 opened - 0
OS-level lock for choosing blocks
#453 opened - 6
- 1
Make client-side macOS support
#449 opened - 0
- 0
API reference
#446 opened - 1
Increase file limit automatically
#444 opened - 1
GGML + petals
#439 opened - 2
- 1
- 1
- 0
- 0
- 5
- 4
Problem running petals on virtual CPU
#421 opened - 3
Request to Enable Git Discussions Feature
#420 opened - 4
Conversational Agent with Llama-2-70b-chat-hf
#419 opened - 2
- 1
how to get only embeddings?
#416 opened - 1
Remember Chat History
#415 opened - 1
How to get faster inference?
#414 opened - 2
- 1
[feature] add privacy
#412 opened - 0
- 6
- 2
- 7
Add windows support
#400 opened - 3
Petals Error: GPU is not available
#398 opened - 2
Cuda 11.8 supported?
#395 opened