Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[DataCap Application] Public Datasets -- Digital Earth Africa #3

Closed
1 of 2 tasks
MaleiWong opened this issue Mar 26, 2024 · 16 comments
Closed
1 of 2 tasks

[DataCap Application] Public Datasets -- Digital Earth Africa #3

MaleiWong opened this issue Mar 26, 2024 · 16 comments

Comments

@MaleiWong
Copy link

Version

1

DataCap Applicant

DataFortress

Project ID

001

Data Owner Name

Environment

Data Owner Country/Region

South Africa

Data Owner Industry

Environment

Website

https://www.digitalearthafrica.org/

Social Media Handle

Twitter: https://twitter.com/DEarthAfrica Youtube: https://www.youtube.com/channel/UCBasD3Dz-TdQTRoyG30kVVw

Social Media Type

Slack

What is your role related to the dataset

Data Preparer

Total amount of DataCap being requested

10

Unit for total amount of DataCap being requested

PiB

Expected size of single dataset (one copy)

1

Unit for expected size of single dataset

PiB

Number of replicas to store

10

Weekly allocation of DataCap requested

1

Unit for weekly allocation of DataCap requested

PiB

On-chain address for first allocation

f1bd4ibu3pugyrn6gaxw2t6ebaskr7oj44l6bvzky

Data Type of Application

Public, Open Dataset (Research/Non-Profit)

Custom multisig

  • Use Custom Multisig

Identifier

No response

Share a brief history of your project and organization

DataFortress boasts a team of 20 technical experts who are dedicated to exploring the IPFS and Filecoin domains and possess extensive experience in distributed storage and computing. Our team has a deep understanding of the key technologies of IPFS and Filecoin and has accumulated a wealth of experience in the field of DC data downloading and distribution. We are capable of providing comprehensive technical solutions to clients.

Is this project associated with other projects/ecosystem stakeholders?

No

If answered yes, what are the other projects/ecosystem stakeholders

No response

Describe the data being stored onto Filecoin

Landsat scenes and metadata. DE Africa provides data from Landsat 5, 7 and 8 satellites, including historical observations dating back to late 1980s and regularly updated new acquisitions.

Where was the data currently stored in this dataset sourced from

AWS Cloud

If you answered "Other" in the previous question, enter the details here

No response

If you are a data preparer. What is your location (Country/Region)

Hong Kong

If you are a data preparer, how will the data be prepared? Please include tooling used and technical details?

No response

If you are not preparing the data, who will prepare the data? (Provide name and business)

No response

Has this dataset been stored on the Filecoin network before? If so, please explain and make the case why you would like to store this dataset again to the network. Provide details on preparation and/or SP distribution.

No response

Please share a sample of the data

s3://deafrica-landsat/
s3://deafrica-landsat-inventory/

Confirm that this is a public dataset that can be retrieved by anyone on the Network

  • I confirm

If you chose not to confirm, what was the reason

No response

What is the expected retrieval frequency for this data

Yearly

For how long do you plan to keep this dataset stored on Filecoin

1.5 to 2 years

In which geographies do you plan on making storage deals

Greater China, Asia other than Greater China

How will you be distributing your data to storage providers

Cloud storage (i.e. S3), HTTP or FTP server, Shipping hard drives

How did you find your storage providers

Big Data Exchange, Partners

If you answered "Others" in the previous question, what is the tool or platform you used

No response

Please list the provider IDs and location of the storage providers you will be working with.

1.f01025366, ShanDong-QingDao
2.f02940074, JiangXi-FuZhou
3.f02951064, ZheJiang-JinHua
4.f02984282, HeNan-ZhengZhou

How do you plan to make deals to your storage providers

Boost client

If you answered "Others/custom tool" in the previous question, enter the details here

No response

Can you confirm that you will follow the Fil+ guideline

Yes

Copy link
Contributor

datacap-bot bot commented Mar 26, 2024

Application is waiting for governance review

@nicelove666
Copy link
Owner

@MaleiWong I am very happy to see your application. Please answer these questions: If you are a data preparer, how will the data be prepared? Please include tooling used and technical details?

@MaleiWong
Copy link
Author

@MaleiWong我很高兴看到你的申请。请回答以下问题: 如果您是数据准备者,将如何准备数据?请包括使用的工具和技术细节?

We carry out secondary development based on https://github.com/karust/gogetcrawl. After downloading a batch of data from Amazon Web Services S3, it will automatically process the packaging and generate the car file. As long as the same download parameters are set, it can be downloaded and Generate car files with the same piececid to ensure that different SPs can download and encapsulate the same file to ensure that the data backup meets the requirements. After the car file is generated, a corresponding unit data will be generated at the same time, including filename, data_cid, piece_cid, piece_size and other information and written into the csv file. The csv file is the datacap distribution file that will be used later, and the file name is the downloaded The marker name and starting position of the dataset. Based on the metadata file, you can tell which portion of the total data set is saved by each LDN. The data download can be executed directly on the SP's device, or we can mail it to the SP through the hard disk after processing.

Copy link
Contributor

datacap-bot bot commented Mar 26, 2024

Datacap Request Trigger

Total DataCap requested

10PiB

Expected weekly DataCap usage rate

1PiB

Client address

f1bd4ibu3pugyrn6gaxw2t6ebaskr7oj44l6bvzky

Copy link
Contributor

datacap-bot bot commented Mar 26, 2024

Application is ready to sign

Repository owner deleted a comment from datacap-bot bot Mar 26, 2024
Copy link
Contributor

datacap-bot bot commented Mar 26, 2024

Request Approved

Your Datacap Allocation Request has been approved by the Notary

Message sent to Filecoin Network

bafy2bzacecu5dhudxxrdlh5ynnw36jijpjsgxg2wy3mrhmfx22rhflflzellu

Address

f1bd4ibu3pugyrn6gaxw2t6ebaskr7oj44l6bvzky

Datacap Allocated

1PiB

Signer Address

f16hmuu3w247dkkhsrbbcbeqbugmpjbxpkrpcdatq

Id

46aa3b66-3253-43f9-bc5a-8aae89f4402e

You can check the status of the message here: https://filfox.info/en/message/bafy2bzacecu5dhudxxrdlh5ynnw36jijpjsgxg2wy3mrhmfx22rhflflzellu

Copy link
Contributor

datacap-bot bot commented Mar 26, 2024

Application is Granted

@nicelove666
Copy link
Owner

I will keep an eye on your distribution plan, please stay compliant

@MaleiWong
Copy link
Author

1.f01025366, ShanDong-QingDao
2.f02940074, JiangXi-FuZhou
3.f02951064, ZheJiang-JinHua
4.f02984282, HeNan-ZhengZhou
5.f0122215, ShanDong-QingDao
6. f0427989, ShanDong-QingDao
7. f02942808, GuangDong-ShenZhen
8. f02894875, JiangXi-FuZhou
9. f02370792, GuangDong-DongGuan
We have currently identified the above SPs. Since the total data size and packaging speed that each SP is prepared to store are different, in order to ensure that there are enough copies of the data, we will contact new SPs to participate in the future.

@nicelove666
Copy link
Owner

@MaleiWong The amount you apply for every week is too large. Why do you apply for 1PIB? Do you have data processing experience? Can you comply with the rules of FIL+? How to ensure that DC will not be abused?

@MaleiWong
Copy link
Author

@MaleiWong The amount you apply for every week is too large. Why do you apply for 1PIB? Do you have data processing experience? Can you comply with the rules of FIL+? How to ensure that DC will not be abused?

DataFortress boasts a team of 20 technical experts who are dedicated to exploring the IPFS and Filecoin domains and possess extensive experience in distributed storage and computing. 2 years of experience in DC data processing.

In v3.1 we wrote weekly requests as 1PiB. But there was only 512TiB in the first round. Therefore, this time we also write the weekly application as 1PiB. We will adjust weekly applications to 200-300TiB in the future.

Compliance is very important. The core of DC is to bring precious data storage assets to Filecoin, which is what we have been pursuing. We have written down our work process in detail in the application form, and you can also download our data to view at any time. We hope to be treated with kindness, thank you.

@nicelove666
Copy link
Owner

Sorry, the quota of 1Pib in the first round is still too much. Although I understand your statement, you think that writing 1Pib in the first round may only be 512Tib, so the first round you want to get is 512T.

I suggest you close this ldn and reapply. The weekly application is recommended not to exceed 300T.

Copy link
Contributor

datacap-bot bot commented Apr 3, 2024

Application is in Refill

@datacap-bot datacap-bot bot added Refill and removed granted labels Apr 3, 2024
Copy link
Contributor

datacap-bot bot commented Apr 3, 2024

DataCap Allocation requested

Multisig Notary address

Client address

f1bd4ibu3pugyrn6gaxw2t6ebaskr7oj44l6bvzky

DataCap allocation requested

512TiB

Id

6633d351-db72-4d3b-b98f-ca2c1507e6c4

Copy link
Contributor

datacap-bot bot commented Apr 3, 2024

Application is ready to sign

Copy link
Contributor

datacap-bot bot commented Apr 22, 2024

DataCap and CID Checker Report Summary1

Storage Provider Distribution

✔️ Storage provider distribution looks healthy.

Deal Data Replication

✔️ Data replication looks healthy.

Deal Data Shared with other Clients2

✔️ No CID sharing has been observed.

Full report

Click here to view the CID Checker report.

Footnotes

  1. To manually trigger this report, add a comment with text checker:manualTrigger

  2. To manually trigger this report with deals from other related addresses, add a comment with text checker:manualTrigger <other_address_1> <other_address_2> ...

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants