Automated farming and account management bot for Datahive.ai platform.
๐จ CRITICAL UPDATE (v1.1.0): Mandatory Database Migration Required! If updating, you MUST run
python migrate.pybefore starting the bot. See Database Migration section below.
โ ๏ธ Note: 10% of registrations support the developer through referral codes. Thank you for using this free software!
| Software | Minimum Version | Recommended |
|---|---|---|
| Python | 3.9 | 3.11+ |
| PostgreSQL | 16 | 16+ |
| Resource | Notes |
|---|---|
| RAM | Depends on accounts/threads. Start with 1GB |
| Disk | 500MB+ (logs can grow) |
| Proxies | HTTP/SOCKS5. Test which work for your region |
# Update & install packages
sudo apt update && sudo apt upgrade -y
sudo apt install -y git curl wget python3 python3-pip python3-venv
# PostgreSQL 16
sudo sh -c 'echo "deb http://apt.postgresql.org/pub/repos/apt $(lsb_release -cs)-pgdg main" > /etc/apt/sources.list.d/pgdg.list'
wget --quiet -O - https://www.postgresql.org/media/keys/ACCC4CF8.asc | sudo apt-key add -
sudo apt update && sudo apt install -y postgresql-16
sudo systemctl start postgresql && sudo systemctl enable postgresqlCreate database:
sudo -u postgres psqlCREATE USER <YOUR_USER> WITH PASSWORD '<YOUR_PASSWORD>';
CREATE DATABASE <YOUR_DB> OWNER <YOUR_USER>;
GRANT ALL PRIVILEGES ON DATABASE <YOUR_DB> TO <YOUR_USER>;
\qInstall bot:
git clone https://github.com/BRN-SLP/Datahive-Bot.git
cd Datahive-Bot
python3 -m venv venv && source venv/bin/activate
pip3 install --upgrade pip
pip3 install -r requirements.txtSee VPS_SETUP.md for detailed instructions.
If you are updating from an older version, you MUST run the migration script to add the last_initialized_at column to your database. This is required for the new daily initialization logic.
# Activate your environment
source venv/bin/activate
# Run the migration
python migrate.pyNote: If you are installing for the first time, this is not required (the table is created correctly automatically), but running it won't hurt.
Edit config/config.yaml:
application_settings:
database_url: "postgres://<USER>:<PASSWORD>@localhost:5432/<DATABASE>"threads:
registration: 1 # MUST BE 1 (rate limit protection)
farming: 100 # Parallel farming threadsmultiprocess_farming:
enabled: true # Enable multiprocess mode
max_processes: 3 # Number of processes (0 = auto)farm_settings:
max_devices_per_batch: 600
max_concurrent_tasks: 250
device_task_timeout: 60referral_code_settings:
source: "db" # "db" = random from database
# "file" = random from referral_codes.txt
# "static" = use static_referral_code
static_referral_code: "" # Used when source: "static"delay_before_start:
min: 60
max: 180
retry:
delay_seconds: 10
max_registration_attempts: 5
proxy_rotation: true
proxy_rotation_after_timeouts: 3 # Rotate proxy after N consecutive timeoutsredirect_settings:
enable: false
email: "your_email@gmail.com"
password: "your_app_password"
imap_server: "imap.gmail.com"imap_settings:
use_proxy_for_imap: false
timeout: 30
servers:
gmail.com: imap.gmail.com
icloud.com: imap.mail.me.com
# ... more serversLocated in config/data/:
| File | Format |
|---|---|
login_accounts.txt |
email:password (one per line) |
farm_accounts.txt |
email (one per line, empty = all) |
proxies.txt |
http://user:pass@host:port |
referral_codes.txt |
code (one per line) |
export_stats_accounts.txt |
email (one per line, empty = all) |
wallets.txt |
private_key (Base58, one per line) |
export.zip |
Apple Health Data Archive |
The bot now supports automated wallet binding and lucrative missions (up to 20,500 points per account).
The bot intercepts the Datahive wallet payload to securely sign transactions headless-ly via solders, without needing a browser or Phantom extension.
Setup:
- Export the Private Key (in Base58 format) from your Phantom/Solflare wallets.
- Paste them into
config/data/wallets.txt(one per line). - The bot will automatically pull a random wallet from this list for each bound account.
The bot uses a Memory-Safe Streaming XML Mutator to randomize your biometric data (heartbeat timestamps, step counts, etc.) on the fly. This ensures Datahive never detects duplicate uploads while keeping memory usage under 20MB. Setup:
- Export your real Apple Health data from your iPhone (Profile -> Export All Health Data).
- Rename the resulting file to exactly
export.zip. - Place this file inside the
config/data/directory.
You do not need to install the Datahive Extension or log into Amazon.
The bot actively accesses your local database to extract the device_id from your long-running farming workers. It then emulates the backend Extension API (intercepting ping and job payload queues) to convince Datahive that your worker is actively crawling Amazon, netting you the 500 points anonymously.
source venv/bin/activate
python main.pyMenu:
Login accounts- Register new accountsFarm accounts- Start farmingExport stats- Export to CSV (results/stats/)Clear proxies- Clear proxy assignmentsExit
For background running, see VPS_SETUP.md (tmux/screen).
| Problem | Solution |
|---|---|
| Database error | Check PostgreSQL is running |
| Rate limit 429 | Add more proxies |
| OTP not found | Check IMAP settings |
Made with โค๏ธ by BRN.SLP
Real parallelism = min(farming_threads ร max_processes, max_concurrent_tasks)
| Your Setup | Formula | Result |
|---|---|---|
| 500 accounts ร 2 devices | = 1000 total devices | |
| max_concurrent_tasks: 250 | 1000 รท 250 | 4 iterations |
| VPS Specs | farming | max_processes | max_concurrent_tasks | Best for |
|---|---|---|---|---|
| 1 CPU / 1GB | 50 | 1 | 50 | Up to 200 accounts |
| 2 CPU / 2GB | 100 | 2 | 150 | Up to 500 accounts |
| 4 CPU / 4GB | 100 | 3 | 300 | Up to 1000 accounts |
| 8+ CPU / 8GB+ | 150 | 4-6 | 500 | 1000+ accounts |
threads:
registration: 1 # Always 1!
farming: 100
multiprocess_farming:
enabled: true
max_processes: 3 # 3 parallel processes
farm_settings:
max_devices_per_batch: 600
max_concurrent_tasks: 300 # 100 ร 3 = 300
device_task_timeout: 60Result: 100 threads ร 3 processes = 300 parallel tasks
โ ๏ธ Important:registrationmust ALWAYS be1with delay โฅ60 seconds to avoid rate limits!