poloniex-agrees-to-settle-with-the-sec-for-more-than-10m.png

Zakaj in kako bi se morali naučiti produktivne znanosti o podatkih?

Izvorno vozlišče: 1858780

Zlati blogZakaj in kako bi se morali naučiti produktivne znanosti o podatkih?

What is Productive Data Science and what are some of its components?




Vir slikpixabay (Brezplačna slika)

Efficiency in data science workflow

 
Data science and machine learning can be practiced with varying degrees of efficiency and productivity. Irrespective of the application area or specialization, a data scientist — beginner or seasoned professional — should strive to enhance his/her efficiency at all aspects of typical data science tasks,

  • statistical analysis,
  • visualization,
  • model selection, feature engineering,
  • code quality testing, modularization,
  • parallel processing,
  • easy web-app deployment



Vir slikpixabay (Brezplačna slika)

 

This means performing all of these tasks,

  • at higher speed
  • with faster debugging
  • in a synchronized manner
  • by taking full advantage of any and all available hardware resources

What should you expect to learn in this process?

 
Let’s imagine somebody is teaching a “Produktivna podatkovna znanost” course or writing a book about it — using Python as the language framework. What should the typical expectations be from such a course or book?



Vir slikpixabay (Brezplačna slika)

 

The course/book should be intended for those who wish to leapfrog beyond the standard way of performing data science and machine learning tasks and utilize the full spectrum of the Python data science ecosystem for a much higher level of productivity.

Readers should be taught how to look out for inefficiencies and bottlenecks in the standard process and how to think beyond the box.

Automation of repetitive data science tasks is a key mindset that the readers will develop from reading this book. In many cases, they will also learn how to extend the existing coding practice to handle larger datasets with high efficiency with the help of advanced software tools that already exist in the Python ecosystem but are not taught in any standard data science.

This should not be a regular Python cookbook teaching standard libraries like Numpy or Pandas.

Rather, it should focus on useful techniques such as how to measure the memory footprint and execution speed of ML models, preizkus kakovosti a data science pipeline, modulirati a data science pipeline for app development, etc. It should also cover Python libraries which come in very handy for avtomatizacija in pospeševanje the day-to-day tasks of any data scientist.

Furthermore, it should touch upon tools and packages which help a data scientist tackling large and complex datasets in a far more optimal way than what would have been possible by following standard Python data science technology wisdom.

Some specific skills to master

 



Vir slikpixabay (Brezplačna slika)

 

To put things in concrete terms, let us summarize some specific skills to master for learning and practicing Produktivna podatkovna znanost. I have also tried to throw in the links to some representative articles to go with each skill as a reference.

  1. kako write fast and efficient code for data science/ML and how to measure their speed and efficiency (glejte ta članek)
  2. How to build modularized and expressive data science pipelines to improve productivity (glejte ta članek)
  3. How to write testing modules for data science and ML models (glejte ta članek)
  4. How to handle large and complex datasets efficiently (which would have been difficult with traditional DS tools)
  5. How to fully utilize GPU and multi-core processors for all kinds of data science and analytics tasks, and not just for specialized deep learning modeling (glejte ta članek)
  6. How to whip up quick GUI apps for the demo of a data science/ML idea or model tuning (glejte ta članek), or how to easily (and quickly) deploy ML models and data analysis code at an app-level (glejte ta članek)

An ideal book on this topic will…

 



Vir slikpixabay (Brezplačna slika)

 

  1. Teach how to look out for inefficiencies and bottlenecks in the standard data science code and how to think beyond the box to solve those problems.
  2. Teach how to write modularized, efficient data analysis and machine learning code to improve productivity in a variety of situations — exploratory data analysis, visualization, deep learning, etc.
  3. Cover a wide range of side topics such as software testing, module development, GUI programiranjeML model deployment as web-app, which are invaluable skillsets for budding data scientists to possess and which are hard to find collectively in any one standard data science book.
  4. Cover parallel computing (e.g., Dask, Ray), scalability (e.g, Vaex, Modin), and GPU-powered data science stack (HITRI) with hands-on examples.
  5. Expose and guide the readers to a larger and ever-expanding Python ecosystem of data science tools that are connected to the broader aspects of inženiring programske opreme and production-level deployment.

A concrete example: GPU-powered and distributed data science

 
While the use of GPUs and distributed computing is widely discussed in the academic and business circles for core AI/ML tasks, they have found less coverage in their utility for regular data science and data engineering tasks. However, using GPUs for regular day-to-day statistical analyses or other data science tasks can go a long way towards becoming the proverbial “Productive Data Scientist".

Na primer, RAPIDS zbirka programskih knjižnic in API-jev vam – navadnemu podatkovnemu znanstveniku (in ne nujno izvajalcu poglobljenega učenja) – možnost in prilagodljivost za izvajanje cevovodi podatkovne znanosti in analitike od konca do konca v celoti na grafičnih procesorjih.



Vir slik: Avtor je ustvaril kolaž

 

When used even with a modest GPU, these libraries show remarkable improvement in speed over their regular Python counterparts. Naturally, we should embrace these whenever we can for Produktivna podatkovna znanost potek dela


 

Similarly, there are excellent open-source opportunities to go beyond the limits of the single-core nature of Python language and embrace the parallel computing paradigm without shifting away from the quintessential data scientist persona.



Vir slik: Avtor je ustvaril kolaž

Povzetek

 
We discussed the utilities and core components of a Produktivna podatkovna znanost workflow. We imagined what an ideal course or book on this topic would offer to the readers. We touched upon some concrete examples and illustrated the benefits. Some related resources were also provided in the context of skills to master.

Lahko preverite avtorjevo GitHub odlagališča za kodo, ideje in vire v strojnem učenju in znanosti o podatkih. Če ste, podobno kot jaz, navdušeni nad AI / strojnim učenjem / znanosti o podatkih, vas prosimo, dodaj me na LinkedIn or sledi mi na tviterju.

 
prvotni. Poročeno z dovoljenjem.

Povezano:

Source: https://www.kdnuggets.com/2021/07/learn-productive-data-science.html

Časovni žig:

Več od KDnuggets