Bittensor Desearch (Subnet 22):
Welcome to Desearch, the AI-powered search engine built on Bittensor. Designed for the Bittensor community and general internet users, Desearch delivers an unbiased and verifiable search experience. Through our API, developers and AI builders are empowered to integrate AI search capabilities into their products, with access to metadata from platforms like X, Reddit, Arxiv and general web search.
- AI-powered Analysis: Utilizes decentralized AI models to deliver relevant, contextual, and unfiltered search results.
- Real-time Access to Diverse Data Sources: Access metadata from platforms like X, Reddit, Arxiv, and broader web data.
- Sentiment and Metadata Analysis: Determines the emotional tone of social posts while analyzing key metadata to provide a comprehensive understanding of public sentiment.
- Time-efficient: Minimizes manual data sorting, saving valuable research time.
- User-friendly Design: Suitable for both beginners and experts.
- Decentralized Platform: Built on the Bittensor network, ensures unbiased and highly relevant search results through decentralization.
- Customizability: Tailors data analysis to meet specific user requirements.
- Versatility: Applicable for diverse research fields, from market analysis to academic studies.
- Community-driven Innovation: Built and optimized by a decentralized network of Bittensor miners, validators, and users for continuous search result enhancement.
Requirements: Python 3.10 or higher
- Clone the repository:
git clone https://github.com/Datura-ai/smart-scrape.git
- Install the requirements:
cd smart-scrape python -m pip install -r requirements.txt python -m pip install -e .
Before running a miner or validator, ensure to:
For setting up the necessary environment variables for your miner or validator, please refer to the Environment Variables Guide.
python -m neurons/miners/miner.py
--netuid 22
--subtensor.network finney
--wallet.name <your miner wallet>
--wallet.hotkey <your validator hotkey>
--axon.port 14000
These validators are designed to run and update themselves automatically. To run a validator, follow these steps:
- Install this repository, you can do so by following the steps outlined in the installation section.
- Install Weights and Biases and run
wandb login
within this repository. This will initialize Weights and Biases, enabling you to view KPIs and Metrics on your validator. (Strongly recommended to help the network improve from data sharing) - Install PM2 and the
jq
package on your system. On Linux:On Mac OSsudo apt update && sudo apt install jq && sudo apt install npm && sudo npm install pm2 -g && pm2 update
brew update && brew install jq && brew install npm && sudo npm install pm2 -g && pm2 update
- Run the
run.sh
script which will handle running your validator and pulling the latest updates as they are issued.pm2 start run.sh --name smart_scrape_validators_autoupdate -- --wallet.name <your-wallet-name> --wallet.hotkey <your-wallet-hot-key>
This will run two PM2 process: one for the validator which is called smart_scrape_validators_main_process
by default (you can change this in run.sh
), and one for the run.sh script (in step 4, we named it smart_scrape_validators_autoupdate
). The script will check for updates every 30 minutes, if there is an update then it will pull it, install it, restart smart_scrape_validators_main_process
and then restart itself.
For step-by-step guidance on setting up and running a miner, validator, or operating on the testnet or mainnet, refer to the following guides:
- Miner Setup
- Validator Setup
- Testnet Operations
- Mainnet Operations
- Setting Up and Running the Web Application with Validator Integration
The text prompting validator sends data to wandb, allowing real-time monitoring with key metrics like:
- Gating model loss
- Hardware usage
- Forward pass time
- Block duration
Data is publicly available at this link. Note that data from anonymous users is deleted after 7 days.