You can not select more than 25 topics Topics must start with a chinese character,a letter or number, can include dashes ('-') and can be up to 35 characters long.

get_start_examples.rst 3.6 kB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263
  1. .. _examples:
  2. ================================
  3. Experiments & Get Start Examples
  4. ================================
  5. This chapter will introduce related experiments to illustrate the search and reuse performance of our learnware system.
  6. ================
  7. Environment
  8. ================
  9. For all experiments, we used a single linux server. Details on the specifications are listed in the table below. All processors were used for training and evaluating.
  10. ==================== ==================== ===============================
  11. System GPU CPU
  12. ==================== ==================== ===============================
  13. Ubuntu 20.04.4 LTS Nvidia Tesla V100S Intel(R) Xeon(R) Gold 6240R
  14. ==================== ==================== ===============================
  15. ================
  16. Experiments
  17. ================
  18. Datasets
  19. ================
  20. We designed experiments on three publicly available datasets, namely `Prediction Future Sales (PFS) <https://www.kaggle.com/c/competitive-data-science-predict-future-sales/data>`_,
  21. `M5 Forecasting (M5) <https://www.kaggle.com/competitions/m5-forecasting-accuracy/data>`_ and `CIFAR 10 <https://www.cs.toronto.edu/~kriz/cifar.html>`_.
  22. For the two sales forecasting data sets of PFS and M5, we divide the user data according to different stores, and train the Ridge model and LightGBM model on the corresponding data respectively.
  23. For the CIFAR10 image classification task, we first randomly pick 6 to 10 categories, and randomly select 800 to 2000 samples from each category from the categories corresponding to the training set, constituting a total of 50 different uploaders.
  24. For test users, we first randomly pick 3 to 6 categories, and randomly select 150 to 350 samples from each category from the corresponding categories from the test set, constituting a total of 20 different users.
  25. We tested the efficiency of the specification generation and the accuracy of the search and reuse model respectively.
  26. The evaluation index on PFS and M5 data is RMSE, and the evaluation index on CIFAR10 classification task is classification accuracy
  27. Results
  28. ================
  29. The time-consuming specification generation is shown in the table below:
  30. ==================== ==================== =================================
  31. Dataset Data Dimensions Specification Generation Time (s)
  32. ==================== ==================== =================================
  33. PFS
  34. M5
  35. CIFAR10 9000*3*32*32 7~10
  36. ==================== ==================== =================================
  37. The accuracy of search and reuse is shown in the table below:
  38. ==================== ==================== ================================= =================================
  39. Dataset Top-1 Performance Job Selector Reuse Average Ensemble Reuse
  40. ==================== ==================== ================================= =================================
  41. PFS
  42. M5
  43. CIFAR10 0.619 +/- 0.138 0.585 +/- 0.056 0.715 +/- 0.075
  44. ==================== ==================== ================================= =================================
  45. =========================
  46. Get Start Examples
  47. =========================
  48. Examples for `PFS, M5` and `CIFAR10` are available at [xxx]. You can run { main.py } directly to reproduce related experiments.
  49. The test code is mainly composed of three parts, namely data preparation (optional), specification generation and market construction, and search test.
  50. You can load data prepared by as and skip the data preparation step.

基于学件范式,全流程地支持学件上传、检测、组织、查搜、部署和复用等功能。同时,该仓库作为北冥坞系统的引擎,支撑北冥坞系统的核心功能。