Skip to content

🧩 fundamentals of software development course | IT conferences crawler frontend

License

Notifications You must be signed in to change notification settings

fit-hcmus-k21/conferences-crawler-fe

Repository files navigation

🧩 IT Conferences Crawler Project

Objective:

  • The main goal of this project is to build a data crawler system to collect information about reputable information technology conferences worldwide.
  • This information may include conference names, deadlines, dates, locations, themes, and additional details such as speaker lists, accepted papers, and registration information.

Suggested Steps:

1. Plan and Design the System:

  • Identify a list of websites or data sources from which you want to gather information.
  • Design the structure of the database to store this information.

2. Develop the Crawler:

  • Build a program or script capable of automatically navigating websites, searching for information about conferences, and extracting data from these web pages.
  • Suggested tools/libraries: Consider using tools like Scrapy or BeautifulSoup in Python, or any language suitable for your team's skills.

3. Data Processing:

  • After collecting the data, process and filter the information to eliminate unnecessary data and ensure data accuracy.

4. Store Data:

  • Store the collected data in a database or file for future use.

5. User Interface Integration (Optional):

  • If necessary, build a user interface to interact with the collected data.

Challenges and Difficulties:

  • Some websites may have anti-crawling mechanisms or limitations on access speed.
  • Data on websites may change frequently, requiring regular updates.
  • Data may not be presented in an easily analyzable format.

Applications:

  • This project can be beneficial for academic researchers, conference organizing entities, or individuals interested in tracking global conference events.

Note:

  • Data collection from websites should adhere to copyright regulations and website policies. The use of collected information may need to comply with legal regulations and the specific policies of each website.

References:

  • CCFDDL
  • LIX Polytechnique
  • Link to crawler with scrapy: go here

About

🧩 fundamentals of software development course | IT conferences crawler frontend

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published