robot_arm/requirements.txt
m1ngsama 9e672ce637 fix(voice_main): pass actual current_pos to execute_pick/execute_lift; add requirements.txt
- Closes #3: execute_pick and execute_lift now accept an optional
  current_pos parameter. Callers (execute_command) pass self.current_pos
  so move_line starts from the actual arm position rather than a hardcoded
  [120, 0, 60] default. Falls back to rest position when None.
- Closes #4: add requirements.txt with correct package names.
  Note: README listed 'openai-whisper' but the code imports faster_whisper;
  requirements.txt uses the correct 'faster-whisper' package name.
- Replace all `servo_buffer = []` assignments with `.clear()` to stay
  compatible with the deque introduced in PR #8 (#6 follow-up).
- Partial #5: translate Chinese print/comment strings to English in
  execute_pick, execute_lift, and execute_command.
2026-02-20 20:26:30 +08:00

33 lines
898 B
Text

# Robot Arm Voice Control System - Dependencies
# Python 3.10+ required
#
# GPU acceleration (CUDA 11.8 / 12.x) is strongly recommended.
# Install PyTorch with CUDA from: https://pytorch.org/get-started/locally/
# Example (CUDA 12.1):
# pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu121
# --- Core scientific stack ---
numpy>=1.24
scipy>=1.11
# --- Serial communication ---
pyserial>=3.5
# --- Computer vision ---
opencv-python>=4.8
ultralytics>=8.0 # YOLOv8
# --- Audio ---
sounddevice>=0.4.6
# --- Speech recognition ---
faster-whisper>=1.0.0 # NOTE: this is faster-whisper, NOT openai-whisper
# --- LLM inference ---
torch>=2.0 # install with CUDA wheels from pytorch.org
transformers>=4.40
accelerate>=0.27
peft>=0.10 # LoRA adapter loading
# --- Visualization (arm_main.py demo) ---
matplotlib>=3.7