Skip to content

vudat1710/fashion_engine

Repository files navigation

Fashion search engine

Posts and images retrieval engine for e-commerce data on multiple platforms on fashion domain. Images are restricted as single type product or only a model wearing a product in an image. Data in this project is crawled from 3 e-commerce platforms: Tiki.vn, Shopee.vn, and Sendo.vn. Crawling code is available at ./ecommerce_crawler. This crawling tool is created using scrapy and scrapy-splash. See more at how to run a scrapy tool at this documentation (you need to install scrapy and scrapy-splash first in order to run this code).

Setup environment and data

  • Python 3 and npm preinstalled
  • Move to backend directory then run pip install -r requirements.txt to setup python environment for backend search engine
  • From outer directory, move to fashion directory then run npm install to install packages for frontend search engine.
  • Install Solr on your machine then create 2 new cores named multimedia (for posts data) and multimedia_shops (for shops data). Default port for solr service is localhost:8983
  • From outer directory, run ./solr_script.sh to create schema for our new cores. If you configure another port for Solr service, be sure to change port number in solr_script.sh.
  • All required files include model for our search engine are available on: This sharepoint. Download these following 4 files: all_feat.list, all_images_path.csv, features_resnet50.npy, res50_sz150_best_stage3_export.pkl to your download folder
  • Move to backend/ folder then change all paths in settings.py file corresponding to paths to your downloaded files above.
  • Download images data for this search engine on This sharepoint. Then create relative path from this images folder to public folder in fashion directory by running the following command: ln -s path_to_your_images_folder this_project_folder/fashion/public
  • Download posts and shops data from This sharepoint (2 jsonl files). Create data directory in outer directory then move these downloaded data to this folder
  • Push data to Solr cores: from outer directory, run python -m backend.import_data
  • Now we are good to go

Starting program

  • From outer directory, run python -m backend.run_api to start backend search APIs. Default port for backend services is localhost:5000
  • Move to fashion folder, run npm start to start frontend service. Default port for frontend service is localhost:3000

Using program

We provide 2 services for users:

  • Firstly, you could choose text search type, type in your keywords you want to search and number of results
  • Secondly, you could choose an image (single type product or only a model wearing the product you want to search) and number of results

Our program will return results corresponding to your input data and provide details of each product with a link to original post from the original platform having that post.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 2

  •  
  •