Motif
Motif

工程師一枚 Θώθ

How troublesome is on-chain data analysis | Use Footprint Analytic to save you time

(edited)
Data analysis on the chain is troublesome, make good use of tools to liberate life

foreword

The analysis of data on the chain is very helpful whether you want to understand a project or you are the project party.

Imagine that you want to invest in a certain coin. You may worry about whether the token economy of the coin is good. At this time, you need to have the ability to analyze the actual situation on the chain, not just watch the project party in the DC. Praise each other with your supporters.

If you are the project party, you may have issued NFTs, which sold very well at first, but after a few months, you may be confused about the current status of your project. By analyzing the data on the chain, you can make your project Not just stop at the success of the release, but can make the most appropriate decision based on the follow-up situation.

But this is a very troublesome thing. This article will share the steps and pains you need to go through to do data analysis from scratch, and how to use tools to quickly achieve your goals and avoid these unnecessary troubles.

On-chain data analysis from scratch

Steps in Data Analysis

get information

Source: https://etherscan.io/chartsync/chaindefault


The most troublesome method is to set up a node and use this node to synchronize the historical data of the blockchain. The above picture shows the data size of the current historical data of Ethereum, so at least 1TB of SSD storage space is required, and other Hardware equipment (for detailed hardware requirements, please refer to ). In addition, the synchronization time will vary according to the speed of the hardware equipment. According to experience, it takes about a week to synchronize to the latest block.

Because self-supporting nodes are too painful, some service providers such as Alchemy and Moralis provide you to obtain on-chain data from their nodes, but these nodes have restrictions on obtaining data, and most of the time the free plan is enough. , but it will take more time to wait.

decode data

After obtaining the data, the second step is to clean the data. At this point in time, you will have tens of millions of Transcation data at hand.

The data format and meaning of Transaction are taken from Etherscan

At this time, there is information about the interaction between participants from all over the world and the chain. The next step is to be able to decode the input data (Decode) so that people can understand it. To do this requires having those ABIs for the item you want to decode.

The role of ABIs in contract interactions

Obtaining ABI needs to be provided by the project party. In addition, if the project party has uploaded Etherscan, you can also use the Ethersan api to obtain a large number of them. Decoding different project contracts requires different ABIs. Therefore, if you are curious about the information of the world's top 100 NFTs, then You will need to collect these hundred ABIs. It was another troublesome and arduous journey, but after decoding it, you can see what interactions have occurred in transcation, and it is a big step forward to analyze the data and find insights.

organising materials

After the data is decoded, it needs to be further organized into an easy-to-use table or database.
Since the data of all the transaction data in the world is in hand when the data is obtained, in this step, the data should be organized into a convenient format according to the requirements, and stored in the database for further analysis.

Data Interception, Transformation and Loading (ETL) Image source: https://docs.microsoft.com/en-tw/azure/architecture/data-guide/relational-data/etl

In this step, you can refer to the Ethereum-ETL tool to design the schema according to your needs and organize the data into table data (csv) or relational database. At this point, the data is finally in a usable form!

Data Analysis <br class="smart">At this point, data scientists can't wait to use a variety of statistical and even machine learning tools to extract insights from the collated data, solve their doubts, and test their hypotheses.

Presenting the results <br class="smart">After the analysis is complete, you may want to post your findings on twitter or personal blog, which is indispensable at this time is a beautiful graph that matches your story. Common tools like Matplotlib, plotly can help you.

https://memeprod.sgp1.digitaloceanspaces.com/user-wtf/1601476779885.jpg

The above is a process of analyzing the data on the chain from scratch, but the whole process is full of too much trouble and pain. The whole process takes 3-4 weeks, and the final result may be a tweet, or whatever you asked. The question has lost its timeliness. Next, I will share the refreshing experience of making good use of tools.

On-chain data analysis using tools -- Footprint Analytics

Get data & decode data & organize data

You don't have to do it yourself, Footprint Analytics has it sorted.

On the Query page, there are already a large number of databases and data fields (Schema) available on the left side. This picture shows the data of the opensea nft transaction using the three-stroke code.

date analyzing

In addition to using SQL commands to calculate some statistical values, data analysis also supports opening a jupyterlab through Playground for data analysis, so that you can use the powerful Python suite to your heart's content! Currently this feature is still in beta, looking forward to better use in the future!

Common data analysis suites are built-in

Presenting Results <br class="smart">Building interactive dashboards is easy with powerful built-in features. You can also easily share your results with others, and see what others are interested in and how they started their analysis.

https://www.footprint.network/dashboards

You can click to see other people's dashboards

Epilogue

It really took a lot of time to do on-chain data analysis in the early days. Until recent years, such time-saving and easy-to-use products began to appear, which can really save most of the time in data processing and focus on the data analysis itself.

Next, I will continue to share the tools and methods for analyzing the data on the chain. If you like it, please help me appreciate it or leave a message. Thank you for seeing this.

https://www.footprint.network/@motif






CC BY-NC-ND 2.0

Like my work?
Don't forget to support or like, so I know you are with me..

Loading...

Comment