bigscience-workshop/petals
🌸 Run LLMs at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading
PythonMIT
Pinned issues
Issues
- 0
Feature Request - Support For VLM's
#616 opened by asmit203 - 0
Create a bash script for the docker image
#615 opened by valentimarco - 1
Error using the local llama3.1 model
#609 opened by CzsGit - 1
Performance improving chances in the future
#614 opened by oldcpple - 2
Mac M3 Any Model crashing
#613 opened by andrew-morris-rgs - 0
Question about overlapped serving blocks
#610 opened by jeremyzhangsq - 0
attention_mask = FalconModel._prepare_attn_mask(attention_mask, (batch_size, seq_length), past_length) AttributeError: type object 'FalconModel' has no attribute '_prepare_attn_mask'
#608 opened by peteblank - 1
- 0
multiple gpu support?
#606 opened by pass-pass-pass - 0
Pascal family cards support.
#605 opened by akamaus - 1
Unable to connect to Private Swarm
#589 opened by Rohit-03 - 0
- 2
Meta Llama 3.1
#597 opened by apcameron - 1
NotImplementedError:
#599 opened by pass-pass-pass - 4
Petals doesn't deal with server failure properly
#587 opened by oldcpple - 1
batch processing/parallel processing
#585 opened by oldcpple - 1
Manual management of shards
#573 opened by nrs-status - 1
Donating System Memory?
#586 opened by NavodPeiris - 2
- 1
System_prompt
#581 opened by EvilSumrak2049 - 3
LLama-3-70B support
#578 opened by ELigoP - 0
Error with PyTorch 2.3.0: Missing '_refresh_per_optimizer_state' in 'torch.cuda.amp.grad_scaler'
#576 opened by Priyanshupareek - 13
Error trying to raise Mixtral private swarm server
#569 opened by Qessia - 2
RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cpu and cuda:0! (when checking argument for argument tensors in method wrapper_CUDA_cat)
#575 opened by jmikedupont2 - 5
Latest bump "Bump transformers and accelerate versions (#554)" looks to destroy Falcon support.
#565 opened by STEMBytes - 5
- 0
DynamicCache and Beam Search
#571 opened by artek0chumak - 3
compile to webassembly
#566 opened by Donovoi - 2
- 3
Grok | Mixture-of-Experts | Model Support
#564 opened by tibrezus - 6
Can not use direct server-to-server communication
#550 opened by miaoqijun - 1
- 0
- 1
Reachability Issue for private swarm
#551 opened by VarunJoshi10 - 2
I confirm Petals not working in Google Colab
#534 opened by AIAnytime - 0
Add "Podman" usage to the documentation
#546 opened by metal3d - 3
ImportError: cannot import name 'AutoDistributedModelForCausalLM' from partially initialized module 'petals' (most likely due to a circular import)
#542 opened by lacksfish - 5
Prepull model data on private swarm
#541 opened by wolfganghuse - 2
M1 macOS installation error ("failed to build wheel for hivemind") with mac-native petals
#540 opened by MinaAlmasi - 2
Upstream Changes makes the demo not work
#536 opened by hrQAQ - 7
IndexError: tuple index out of range
#532 opened by jaskirat8 - 0
Error while hosting as provider
#533 opened by filopedraz - 2
text to video generation models ?
#529 opened by scenaristeur - 0
Pyinstaller packaged Petals binary fails to load/download on`AutoDistributedModelForCausalLM.from_pretrained`
#528 opened by biswaroop1547 - 2
support quivr or privateGPT?
#524 opened by lbgws2 - 1
whether past_key_values can be obtained
#526 opened by sa1utyeggs - 1
Add mistral to chat.petals
#527 opened by AmgadHasan - 0
Add pre-commit hook
#522 opened by mahimairaja - 1
Available Models?
#521 opened by jontstaz - 2
Support stable diffusion model
#519 opened by lbgws2