Skip to content

Latest commit

 

History

History
220 lines (144 loc) · 15 KB

README.md

File metadata and controls

220 lines (144 loc) · 15 KB

QuestDK Plugins

This is the official registry of actions supported on Rabbithole.gg & Terminal.

Getting started

  1. Clone the repo
  2. pnpm install
  3. navigate to a package or create a new one

Testing

To run all tests:

pnpm test

To run individual package test:

pnpm test --filter=@rabbitholegg/questdk-plugin-connext

Replace the filter param with the package you're trying to target.

Building Your Project

To build all projects except the @rabbitholegg/questdk-plugin-project, run the following command:

pnpm build

This command utilizes Turbo to optimize the build process, ensuring that only relevant changes trigger rebuilds, significantly speeding up development time.

Linting

Maintain code quality and consistency with our linting commands. To lint all projects except the @rabbitholegg/questdk-plugin-project, you can run:

pnpm lint

For automatic fixes to linting issues, execute:

pnpm lint:fix

These commands help ensure that your code adheres to our coding standards and styles.

Cleaning the Project

Encountering build issues or need to start from scratch? The nuke command is your go-to solution. It cleans your working directory by removing all untracked files and directories, then reinstalls dependencies:

pnpm nuke

This powerful command is ideal for resetting your project to a clean state.

Remember to replace the filter parameter with the specific package you're targeting for individual commands as needed.

Changesets & Publishing

In order to publish you need to make sure that the pull request you're submitting has a changeset. If you don't want to publish this isn't needed. In order to generate a changeset run pnpm changeset, select a change type [major,minor,patch], and draft a small summary of the changeset. Select version based on semantic versioning.

After this all you need to do is push and merge the pull request and the Github Action will handle the process of versioning, and publishing.

Commit Standards

We use Conventional Commits in order to make changelogs and versioning a breeze.

Package Linking

Often times when testing it's necessary to link into the questDK repo. We handle this through package linking.

We recommend using direct linking, not global linking. If the local questDK changes don't seem to be recognized try deleting the file node_modules\.pnpm\@rabbitholegg+questdk [may have some version specific information - delete all of them].

Also remember the package must be built for changes to take effect in this repo. It should not be necessary to re-link after builds.

At times it may be necessary to restart the typescript server when working with a linked package in VSCode.

In order to link the questdk run this from root of this package:

pnpm link path/to/questdk

If you encounter Module not found: Can't resolve 'viem/chains' or another dependency related issue immediatly after linking you most likely need to rebuild the package that has been linked into the host package.

Plugins: an Overview

What Are Quest Plugins?

Quest Plugins are how we translate transaction data into actions. It’s how our platform can take information from smart contracts deployed for any EVM native protocol and quickly parse it into a standardized format. Using this standardized action information we determine if a given user has transactions that meet the specific criteria for a given quest.

For example, writing a plugin for Uniswap that translates a swap action allows users to create, and complete Quests on Uniswap. The plugin is used by our Indexing service to parse transaction data into specific information about that allowing quests to target certain amounts, certain recipients, or even certain tokens.

At this time we only support plugins for swap, mint, and bridge but we’re constantly adding new actions (stake forthcoming). If there’s a specific action you’d like supported please reach out. Some protocol integrations are more complex than others.

Why Quest Plugins

Plugins are necessary because there is no standard for how protocols should describe actions, so implementations vary across protocols. This makes it difficult to understand the type of action being performed just from transaction data. Plugins serve two purposes: integrate protocol actions using a standard schema, and provide an open source registry of curated protocol actions approved for use on Rabbithole and Terminal.

Without a plugin, it’s not possible to create quests for a specific action, since we don’t have an ability to parse the transaction data and understand the actions that are occurring in those transactions. These plugins allow us to store action information in a standardized way. Once the action has been standardized we can use any of these standard elements as criteria for quests on the platform. As we progress, this also unlocks the ability to perform advanced analytics on a given class of actions.

In the example given above, Uniswap uses an extremely specialized pattern for their Swap contract relying on a bytes array for all input data.

Without a plugin to translate this information into a standardized format that includes chainId amountIn or recipient it wouldn’t be possible for us to establish specific success criteria for quests deployed through our terminal. In this particularly obtuse example, all of the inputs are tucked into a bytes array, so without our plugin it’s impossible to know these common attributes that are shared between all swaps.

Once a protocol has a plugin merged into our plugin repo, that protocol will automatically be a valid option for quest creation on our platform through the quest creation terminal.

Creating a plugin

Plugin Creation Overview

Plugin implementation is relatively simple, although it requires a strong understanding of the project you're integrating with. Oftentimes we rely on a projects API to get fresh and consistent information on the project we're integrating with. Each plugin has a user defined pluginID that needs to be added to the plugin registry and should be descriptive of the project you're integrating. We also require functions to return the list of all supported chains [ getSupportedChainIds ], and the list of supported tokens [ getSupportedTokenAddresses ] for each supported chain. The supported chains and tokens are often where a project specific API can come in handy. The most complex aspect of plugin implementation is the bridge mint and swap functions which are used to return a TransactionFilter. For our backend, we store this filter in the quests actionSpec which we later read when applying filters in our indexer. The generally provides a way of finding or filtering out transactions that fulfill the requirements of a quest’s action property.

Steps to Plugin Creation

When developing a plugin the steps are as follows:

  1. Create a new package in packages by duplicating the template folder, following the naming convention of questdk-plugin-<project>
  2. Fill out the appropriate test transaction information in the test-transaction.ts file. If you find an example of a transaction, this tool can help generate the test-transaction data. These transactions are meant to cover all of the edge cases for a given action. Oftentimes the contract, and function signature will be different for these different transaction types. In extreme situations there may be additional edge cases (handling of Matic on Polygon for example) but ensure that every situation where a different contract, or a different function signature is used is fully captured in the file. These are the transactions you should be testing against, and you can allow these different edge cases to influence how the plugin is developed.
// BRIDGE TEST TRANSACTIONS
const BRIDGE_ETH_L1_TO_L2 = {}
const BRIDGE_ETH_L2_TO_L1 = {}
const BRIDGE_TOKEN_L1_TO_L2 = {}
const BRIDGE_TOKEN_L2_TO_L1 = {}
// SWAP TEST TRANSACTIONS
const SWAP_TOKEN_TO_TOKEN = {}
const SWAP_TOKEN_TO_ETH = {}
const SWAP_ETH_TO_TOKEN = {}
// MINT TEST TRANSACTIONS
const MINT = {}
  1. Implement IActionPlugin and export the interface as a top-level package export. In the case that not all actions are implemented, just return PluginActionNotImplementedError:
    1. Implement previously mentioned helper functions [ getSupportedChainIds , getSupportedTokenAddresses ]
    2. Implement any action functions you expect the plugin to support [ swap, mint, bridge ]
  2. Add your plugin by id in the plugin registry questdk-plugin-registry.
    1. id should be listed using kebab-case and any versioning should be appended (uniswap-v2, mirror-world, etc)

How to Implement Action Functions

When implementing the action function, the parameters coming in are the criteria set through the terminal:

const { 
	sourceChainId, 
	destinationChainId, 
	tokenAddress,  // For Native Tokens (ETH, Matic, etc) sometimes this is 0x0 or blank
	amount, 
	recipient
} = bridge

Given these we have to map these supplied expected values against the actual ABI of the transaction we’ll be parsing:

return compressJson({
    chainId: sourceChainId, // The chainId of the source chain
    to: CHAIN_TO_CONTRACT[sourceChainId], // The contract address of the bridge
    input: {
      $abi: ACROSS_BRIDGE_ABI, // The ABI of the bridge contract
      recipient: recipient, // The recipient of the funds
      destinationChainId: destinationChainId, // The chainId of the destination chain
      amount: amount, // The amount of tokens to send
      originToken: tokenAddress, // The token address of the token to send
    }, // The input object is where we'll put the ABI and the parameters
  })

chainId and to both map directly to those params on the transaction object. Any param on the transaction object can be used in your filter (from is often useful). For the input object you need to supply an ABI that contains the function signature of the relevant function. This can be a modified ABI that holds the function signature of multiple contracts, the filter will correctly pull out the one with the right signature, just watch out for poor handling of overloaded functions. The keys of the JSON object (i.e originToken ) needs to key into the expected value based on the parameters passed in from the action (tokenAddress).

Adding To Registry

In the registry index.ts import your new plugin and add it to the plugins object as follows:

  [Project.pluginId]: Project,

Replace 'Project' with the name of your project.

Also remember to add the new repo within the monorepo to the package.json of the registry.

Private Repo

If you're ready to publish, remove the private tag from your package.json file.

Build Process Update

We've made significant updates to our build process to enhance modularity, improve integration times, and set a foundation for reducing the overall package size. Although the package remains larger than our target, these changes pave the way for future optimizations through deduplication and exclusions.

Motivations for the Update

  • Enhanced Portability: The new structure improves the portability of modules, facilitating their use in diverse environments.
  • Foundation for Slimming Down: By breaking down the build process and migrating away from unsupported tools, we've laid the groundwork for future size reductions.
  • Improved Integration and Performance: Anecdotal evidence suggests a smoother and faster integration process, likely due to more efficient tree-shaking capabilities.

New Build Flow

Our build process now consists of three discrete steps, designed to ensure both type safety and build speed:

  1. Type Verification: We start by verifying types and building type definitions. This step ensures the type safety of our code before it moves to transpilation.
  2. Transpilation: The code is then transpiled into unbundled CommonJS (CJS) and ECMAScript Module (ESM) formats using Babel. This step focuses on speed, bypassing the checks performed in the first step.
  3. Bundling: Finally, we bundle the transpiled code into ESM and Universal Module Definition (UMD) formats. This step also includes the minification of each version to reduce the size further.

Additional Updates

  • Third-party SDK's: If a plugin is dependent on a third party SDK, like @zoralabs/protocol-sdk, please ensure that a matching package name regular expression, ie: /@zoralabs/ is added to the plugin's vite.config.js build.rollupOptions.external to prevent unnecessary inclusion in compiled output.
  • Migration Away from Rome: Recognizing that Rome is no longer maintained, we've started to migrate our tooling away from it to ensure future compatibility and support.
  • Utility Script for File Management: To facilitate cleaner updates, a helper script has been added for copying files out to every directory. This approach allows for a more straightforward "delete all and replace" strategy when updating modules.

Remaining Tasks and Considerations

While the new build process marks a significant improvement, some tasks remain, particularly concerning node polyfills:

  • Node Polyfills with Rollup: Issues have arisen around TextDecoder, which is unavailable in certain environments. While we've managed to polyfill node utilities for most packages, exceptions exist for those utilizing Axios due to its reliance on TextDecoder.
  • Web Environment Compatibility: As TextDecoder should be available in web environments, further investigation is needed to selectively polyfill node utilities. This consideration is crucial for web browser deployment, suggesting that webpack may offer a solution for filling these gaps.

We are committed to addressing these challenges and will provide updates as we refine the build process further. Your feedback and contributions are welcome as we continue to improve the efficiency and effectiveness of our package.

Contributing

If you'd like to build a plugin and get support for your protocol on RabbiteHole all you need to do is submit a PR with the finished plugin. Here are some useful tips to assist, and when in doubt please join our discord or reach out by email [[email protected]] for assistance building a plugin.