Dataframe Manipulation Theory And Applications With Python And Tkinter

eBook Download

BOOK EXCERPT:

A DataFrame is a fundamental data structure in pandas, a powerful Python library for data manipulation and analysis, designed to handle two-dimensional, labeled data akin to a spreadsheet or SQL table. It simplifies working with tabular data by supporting various operations like filtering, sorting, grouping, and aggregating. DataFrames are easily created from lists, dictionaries, or NumPy arrays and offer flexible data handling, including managing missing values and performing input/output operations with different file formats. Key features include hierarchical indexing for multi-level grouping, time series functionality, and integration with libraries such as NumPy and Matplotlib. DataFrame manipulation encompasses filtering, sorting, merging, grouping, pivoting, and reshaping data, while also allowing custom functions, handling missing data, and managing data types. Mastering these techniques is crucial for efficient data analysis, ensuring clean, transformed data ready for deeper insights and decision-making. In chapter 2, in the first project, we filter a DataFrame named employee_data, which includes columns like 'Name', 'Department', 'Age', 'Salary', and 'Years_Worked', to find employees in the 'Engineering' department with a salary exceeding $70,000. We create the DataFrame using sample data and apply boolean indexing to achieve this. The boolean masks employee_data['Department'] == 'Engineering' and employee_data['Salary'] > 70000 identify rows meeting each condition. Combining these masks with the & operator filters the DataFrame to include only those rows where both conditions are met, resulting in a subset of employees who fit the criteria. The final output displays this filtered DataFrame. In second project, we filter a DataFrame named sales_data, which includes columns such as 'Product', 'Category', 'Quantity Sold', 'Unit Price', and 'Total Revenue', to find products in the 'Electronics' category with quantities sold exceeding 100. We use boolean indexing to achieve this: sales_data['Category'] == 'Electronics' creates a mask for rows in the 'Electronics' category, while sales_data['Quantity_Sold'] > 100 identifies rows where quantities sold are above 100. By combining these masks with the & operator, we filter the DataFrame to include only rows meeting both conditions. The final output displays this filtered subset of products. In third project, we filter a DataFrame named movie_data, which includes columns such as 'Title', 'Genre', 'Release Year', 'Rating', and 'Box Office Earnings', to find movies released after 2010 with a rating above 8. We use boolean indexing where movie_data['Release_Year'] > 2010 creates a mask for movies released after 2010, and movie_data['Rating'] > 8 identifies movies with ratings higher than 8. By combining these masks with the & operator, we filter the DataFrame to include only the rows meeting both conditions. The final output displays the subset of movies that fit these criteria. The fourth project demonstrates a Tkinter-based GUI application for filtering a sales dataset using Python libraries Tkinter, Pandas, and PandasTable. The application allows users to interact with a table displaying sales data, applying filters based on product category and quantity sold. The filter_data() function updates the table to show only items from the selected category with quantities exceeding the specified value, while the refresh_data() function resets the table to display the original dataset. The GUI includes input fields for category selection and quantity entry, along with buttons for filtering and refreshing. The sales data is initially presented in a PandasTable with a toolbar and status bar. Users interact with the interface, which updates and displays filtered data or the full dataset as needed. The fifth project features a Tkinter GUI application that lets users filter a movie dataset by minimum release year and rating using Python libraries Tkinter, Pandas, and PandasTable. The filter_data() function updates the displayed table based on user inputs, while the refresh_data() function resets it to show the original dataset. The GUI includes fields for entering minimum release year and rating, buttons for filtering and refreshing, and a PandasTable for displaying the data. The application allows for interactive data filtering and visualization, with the table initially populated with sample movie data. In the sixth project, a retail store manager uses a DataFrame containing sales data to identify products that are both popular and profitable. By applying logical operators to filter the DataFrame, the goal is to isolate products that have sold more than 100 units and generated revenue exceeding $5000. This filtering is achieved using the Pandas library in Python, where the & operator combines conditions to select the relevant rows. The resulting DataFrame, which includes only products meeting both criteria, provides insights for decision-making and analysis in retail management. The seventh project involves creating a Tkinter-based GUI application to manage and visualize sales data. The GUI displays data in a table and a bar graph, allowing users to filter products based on minimum quantity sold and total revenue. The application uses pandas for data manipulation, pandastable for table display, and matplotlib for the bar graph. The GUI consists of an input frame for user filters and a display frame for showing the table and graph side by side. Users can update the table and graph by clicking "Filter Data" or reset them to the original data with the "Refresh" button, providing an interactive way to analyze sales performance. In chapter three, the first project demonstrates how to sort synthetic financial data for analysis. The code imports libraries, sets random seeds for reproducibility, and generates data for businesses including revenue and expenses. It then creates a DataFrame with this data, sorts it by monthly revenue in descending order, and saves the sorted DataFrame to an Excel file. This process aids in organizing and analyzing financial data, making it easier to identify top-performing businesses. The second project creates a Tkinter GUI to view and interact with synthetic financial data, displaying monthly revenue and expenses for various businesses. It generates random data, stores it in a DataFrame, and sets up a GUI with two tabs: one for sorting by revenue and another for expenses. Each tab features a table to display the data and a matplotlib plot for visual representation. The GUI allows users to sort and view data dynamically, with alternating row colors for readability and embedded plots for better analysis. The third project generates synthetic unemployment data for 10 regions over 5 years, sets random seeds for reproducibility, and creates a DataFrame with the data. It then sorts the DataFrame alphabetically by region and saves it to an Excel file named "synthetic_unemployment_data.xlsx". Finally, the script prints a confirmation message indicating that the data has been successfully saved. The fourth project generates synthetic unemployment data for 25 regions over a 5-year period and creates a Tkinter GUI for interactive data exploration. The data, organized into a DataFrame and saved to an Excel file, is displayed in a tabbed interface with two views: one sorted by unemployment rate and another by year. Each tab features scrollable tables and corresponding bar charts for visual analysis. The UnemploymentDataGUI class manages the interface, updating tables and graphs dynamically to allow users to explore regional and yearly unemployment variations effectively. The fifth project demonstrates how to concatenate dataframes with synthetic temperature data for various countries. Initially, we generate temperature data for countries like the USA and Canada for each month. Next, we create an additional dataframe with temperature data for other countries such as the UK and Germany. We then concatenate the original and additional dataframes into a single dataframe and save the combined data to an Excel file named combined_temperatures.xlsx. The steps involve generating synthetic data, creating additional dataframes, concatenating them, and exporting the result to Excel. The sixth project demonstrates how to build a Tkinter application to visualize synthetic temperature data. The app features a tabbed interface with tabs for displaying raw data, temperature graphs, and filters. It uses alternating row colors for better readability and includes functionality for filtering data by country and month. Users can view and analyze temperature data across different countries through tables and graphical representations, and apply or reset filters as needed. The seventh project demonstrates how to perform an inner join on two synthetic dataframes: one containing housing details and the other containing owner information. First, synthetic data is generated for houses and their owners. The dataframes are then merged on the common key, HouseID, using an inner join to include only rows with matching keys. Finally, the combined data is saved to an Excel file named combined_housing_data.xlsx. The result is an Excel file that contains details about houses along with their respective owners. The eight project provides an interactive platform for managing and visualizing synthetic housing data. Users can view comprehensive tables, apply filters for location and house type, and analyze house price distributions with Matplotlib plots. The application includes tabs for displaying data, filtering results, and generating visualizations, with functionalities to reset filters, save filtered data to Excel, and ensure a user-friendly experience with alternating row colors in tables and dynamic updates. To demonstrate an outer join on DataFrames with synthetic medical data, in ninth project, we create two DataFrames: one for patient information and another for medical records. We then perform an outer join to ensure all patients and records are included, even if some records don't have corresponding patient data. The code generates synthetic data, performs the outer join using pd.merge() on the PatientID column, and saves the result to an Excel file named outer_join_medical_data.xlsx. This approach provides a comprehensive dataset with complete patient and medical record information. The tenth project involves creating a Tkinter-based desktop application to visualize and interact with synthetic medical data. The application uses an outer join to merge patient and medical record datasets, displaying the comprehensive result in a user-friendly table. Users can filter data by patient ID and condition, view distribution graphs of medical conditions, and save filtered results to an Excel file. The GUI, leveraging Tkinter and Matplotlib, includes tabs for data display, filtering, and graph visualization, providing a robust tool for exploring medical datasets. In chapther four, the first project demonstrates creating and manipulating a synthetic insurance dataset. Using numpy and pandas, the script generates random data including columns for Policyholder, Age, State, Coverage_Type, and Premium. It groups this data by State and Coverage_Type to show basic data segmentation, then saves the dataset to an Excel file for further analysis. The code provides a practical framework for simulating and analyzing insurance data by illustrating the process of data creation, grouping, and storage. The second project demonstrates a Tkinter GUI application designed for analyzing a synthetic insurance dataset. The GUI displays 1,000 records of policyholder data in a scrollable table using the Treeview widget, with options to filter by state and coverage type. Users can save filtered data to an Excel file and generate a bar plot of policy distribution by state, integrated into the Tkinter window using Matplotlib. This application provides interactive tools for data exploration, filtering, exporting, and visualization in a user-friendly interface. The third project focuses on creating, analyzing, and aggregating a large synthetic sales dataset with 10,000 records. This dataset includes salespersons, regions, products, sales amounts, and timestamps, simulating a detailed sales environment. The core task involves grouping the data by region, product, and salesperson to calculate total sales and transaction counts. This aggregated data is saved to an Excel file, providing insights into sales performance and trends, which helps businesses optimize their sales strategies and make informed decisions. The fourth project develops a Tkinter GUI for analyzing synthetic sales data, allowing users to explore raw and aggregated data interactively. The application includes a dual-view setup with raw and aggregated data tables, filtering options for region, product, and salesperson, and visualization features for generating plots. Users can apply filters, view data summaries, save results to Excel, and visualize sales trends by region. The GUI is designed to provide a comprehensive tool for data analysis, visualization, and reporting. The dataset includes 10,000 records with attributes such as salesperson, region, product, sales amount, and date, and is grouped by region, product, and salesperson to aggregate sales data. The fifth project demonstrates how to create and analyze a synthetic transportation dataset. The code generates a large dataset simulating vehicle and route data, including distances traveled and durations. It groups the data by vehicle and route, calculating total and average distances and durations, and then saves these aggregated results to an Excel file. This approach allows for detailed examination of transportation patterns and performance metrics, facilitating reporting and decision-making. The sixth project outlines a Tkinter GUI project for analyzing synthetic transportation data using Python. This GUI, combining Tkinter and Matplotlib, provides a user-friendly interface to inspect and visualize large datasets involving vehicle routes, distances, and durations. It features interactive tables for raw and aggregated data, filter options for vehicle, route, and date, and integrates various plots like histograms and bar charts for data visualization. Users can apply filters, view dynamic updates, and save filtered data to Excel. The goal is to facilitate comprehensive data analysis and enhance decision-making through an intuitive, interactive tool. In chapter five, the first project involves generating and analyzing a synthetic dataset representing gold production across countries, years, and regions. The dataset, created with attributes like country, year, region, and production quantities, simulates complex real-world data for detailed analysis. By using the pivot_table method, the data is transformed to aggregate gold production metrics by country and region over different years, revealing trends and patterns. The results are saved as both original and pivoted datasets in Excel files for easy access and further analysis, aiding in decision-making related to mining and resource management. The second project creates an interactive Tkinter GUI to visualize and interact with a large synthetic dataset on gold production, including details on countries, regions, mines, and yearly production. Using pandas and numpy to generate the dataset, the GUI features multiple tabs for viewing the original data, pivoted data, and various summary statistics, alongside graphical visualizations of gold production trends across countries, regions, and years. The application integrates matplotlib for embedding charts within the Tkinter interface, making it a comprehensive tool for exploring and analyzing the data effectively. The third project demonstrates how to create a synthetic dataset simulating stock prices for multiple companies over 10,000 days, using random number generation to simulate stock prices for AAPL, GOOG, AMZN, MSFT, TSLA, and META. The dataset, initially in a wide format with separate columns for each company's stock prices, is then reshaped to a long format using pd.melt(). This long format, where each row represents a single date, stock, and its price, is often better suited for data analysis and visualization. Finally, both the original and unpivoted DataFrames are saved to separate Excel files for further use. The fourth project involves developing a visually engaging Tkinter GUI to analyze and visualize a synthetic stock dataset. The application handles stock price data for multiple companies, offering users both the original and unpivoted DataFrames, along with summary statistics and graphical representations. The GUI includes tabs for viewing raw and transformed data, statistical summaries, and interactive graphs, utilizing Tkinter's advanced widgets for a polished user experience. Data is saved to Excel files, and Matplotlib charts are integrated for clear data visualization, making the tool useful for both casual and advanced analysis of stock market trends. In chapter six, the first project demonstrates creating a large synthetic road traffic dataset with 10,000 rows using randomization techniques. Fields include Date, Time, Location, Vehicle_Count, Average_Speed, and Incident. Random NaN values are introduced into 10% of the dataset to simulate missing data. The dataset is then cleaned by removing rows with any missing values using dropna(), and the resulting cleaned DataFrame is saved to 'cleaned_large_road_traffic_data.xlsx' for further analysis. The second project creates a Tkinter-based GUI to analyze and visualize a synthetic road traffic dataset. It generates a dataset with 10,000 rows, including fields like date, time, location, vehicle count, average speed, and incidents. Random missing values are introduced and then removed by dropping rows with any NaNs. The GUI features four tabs: one for the original dataset, one for the cleaned dataset, one for summary statistics, and one for distribution graphs. Users can explore data tables with Tkinter's Treeview widget and view visualizations such as histograms and bar charts using Matplotlib, providing a comprehensive tool for data analysis. The third project generates a large synthetic electricity dataset to simulate real-world patterns in electricity consumption, temperature, and pricing. Missing values are introduced and then handled by filling gaps with regional averages for consumption, forward-filling temperature data, and using overall means for pricing. The cleaned dataset is saved to an Excel file, offering a valuable resource for testing data processing methods and developing data analysis algorithms in a controlled environment. The fourth project demonstrates a Tkinter GUI for handling missing data in a synthetic electricity dataset. The application offers a multi-tab interface to analyze electricity consumption data, including features for displaying the original and cleaned DataFrames, summary statistics, distribution graphs, and time-series plots. Users can view raw and processed data, explore statistical summaries, and visualize distributions and trends in electricity consumption, temperature, and pricing over time. The GUI integrates data generation, cleaning, and visualization techniques, providing a comprehensive tool for electricity data analysis.

Product Details :

Genre : Computers
Author : Vivian Siahaan
Publisher : BALIGE PUBLISHING
Release : 2024-08-13
File : 431 Pages
ISBN-13 :


Ultimate Pandas For Data Manipulation And Visualization

eBook Download

BOOK EXCERPT:

TAGLINE Unlock the power of Data Manipulation with Pandas. KEY FEATURES ● Master Pandas from basics to advanced and its data manipulation techniques. ● Visualize data effectively with Matplotlib and explore data efficiently. ● Learn through hands-on examples and practical real-world use cases. DESCRIPTION Unlock the power of Pandas, the essential Python library for data analysis and manipulation. This comprehensive guide takes you from the basics to advanced techniques, ensuring you master every aspect of pandas. You'll start with an introduction to pandas and data analysis, followed by in-depth explorations of pandas Series and DataFrame, the core data structures. Learn essential skills for data cleaning and filtering, and master grouping and aggregation techniques to summarize and analyze your data sets effectively. Discover how to reshape and pivot data, join and merge multiple datasets, and handle time series analysis. Enhance your data analysis with compelling visualizations using Matplotlib, and apply your knowledge in a real-world scenario by analyzing bank customer churn. Through hands-on examples and practical use cases, this book equips you with the tools to clean, filter, aggregate, reshape, merge, and visualize data effectively, transforming it into actionable insights. WHAT WILL YOU LEARN ● Wrangle data efficiently using Pandas' cleaning, filtering, and transformation techniques. ● Unlock hidden patterns with advanced grouping, joining, and merging operations. ● Master time series analysis with Pandas to extract valuable insights from your data. ● Apply Pandas to real-world scenarios like customer churn analysis and financial modeling. ● Unleash the power of data visualization with Matplotlib and craft compelling charts and graphs. ● Enhance your workflow with essential Pandas optimizations and performance tips. WHO IS THIS BOOK FOR? This book is ideal for aspiring data scientists, analysts, and Python enthusiasts looking to enhance their data manipulation skills using Pandas. Familiarity with Python programming basics and a basic understanding of data structures will greatly benefit readers as they delve into the concepts presented in this book. TABLE OF CONTENTS 1. Introduction to Pandas and Data Analysis 2. Pandas Series 3. Pandas DataFrame 4. Data Cleaning with Pandas 5. Data Filtering with Pandas 6. Grouping and Aggregating Data 7. Reshaping and Pivoting in Pandas 8. Joining and Merging Data in Pandas 9. Introduction to Time Series Analysis in Pandas 10. Visualization Using Matplotlib 11. Analyzing Bank Customer Churn Using Pandas Index

Product Details :

Genre : Computers
Author : Tahera Firdose
Publisher : Orange Education Pvt Ltd
Release : 2024-06-10
File : 384 Pages
ISBN-13 : 9788197256240


Artificial Intelligence With Microsoft Power Bi

eBook Download

BOOK EXCERPT:

Advance your Power BI skills by adding AI to your repertoire at a practice level. With this practical book, business-oriented software engineers and developers will learn the terminologies, practices, and strategy necessary to successfully incorporate AI into your business intelligence estate. Jen Stirrup, CEO of AI and BI leadership consultancy Data Relish, and Thomas Weinandy, research economist at Upside, show you how to use data already available to your organization. Springboarding from the skills that you already possess, this book adds AI to your organization's technical capability and expertise with Microsoft Power BI. By using your conceptual knowledge of BI, you'll learn how to choose the right model for your AI work and identify its value and validity. Use Power BI to build a good data model for AI Demystify the AI terminology that you need to know Identify AI project roles, responsibilities, and teams for AI Use AI models, including supervised machine learning techniques Develop and train models in Azure ML for consumption in Power BI Improve your business AI maturity level with Power BI Use the AI feedback loop to help you get started with the next project

Product Details :

Genre : Computers
Author : Jen Stirrup
Publisher : "O'Reilly Media, Inc."
Release : 2024-03-28
File : 499 Pages
ISBN-13 : 9781098112707


Introduction To Data Systems

eBook Download

BOOK EXCERPT:

Encompassing a broad range of forms and sources of data, this textbook introduces data systems through a progressive presentation. Introduction to Data Systems covers data acquisition starting with local files, then progresses to data acquired from relational databases, from REST APIs and through web scraping. It teaches data forms/formats from tidy data to relationally defined sets of tables to hierarchical structure like XML and JSON using data models to convey the structure, operations, and constraints of each data form. The starting point of the book is a foundation in Python programming found in introductory computer science classes or short courses on the language, and so does not require prerequisites of data structures, algorithms, or other courses. This makes the material accessible to students early in their educational career and equips them with understanding and skills that can be applied in computer science, data science/data analytics, and information technology programs as well as for internships and research experiences. This book is accessible to a wide variety of students. By drawing together content normally spread across upper level computer science courses, it offers a single source providing the essentials for data science practitioners. In our increasingly data-centric world, students from all domains will benefit from the “data-aptitude” built by the material in this book.

Product Details :

Genre : Computers
Author : Thomas Bressoud
Publisher : Springer Nature
Release : 2020-12-04
File : 828 Pages
ISBN-13 : 9783030543716


Python For Excel

eBook Download

BOOK EXCERPT:

While Excel remains ubiquitous in the business world, recent Microsoft feedback forums are full of requests to include Python as an Excel scripting language. In fact, it's the top feature requested. What makes this combination so compelling? In this hands-on guide, Felix Zumstein--creator of xlwings, a popular open source package for automating Excel with Python--shows experienced Excel users how to integrate these two worlds efficiently. Excel has added quite a few new capabilities over the past couple of years, but its automation language, VBA, stopped evolving a long time ago. Many Excel power users have already adopted Python for daily automation tasks. This guide gets you started. Use Python without extensive programming knowledge Get started with modern tools, including Jupyter notebooks and Visual Studio code Use pandas to acquire, clean, and analyze data and replace typical Excel calculations Automate tedious tasks like consolidation of Excel workbooks and production of Excel reports Use xlwings to build interactive Excel tools that use Python as a calculation engine Connect Excel to databases and CSV files and fetch data from the internet using Python code Use Python as a single tool to replace VBA, Power Query, and Power Pivot

Product Details :

Genre : Business & Economics
Author : Felix Zumstein
Publisher : "O'Reilly Media, Inc."
Release : 2021-03-04
File : 338 Pages
ISBN-13 : 9781492080978


Architecting Solutions With Sap Business Technology Platform

eBook Download

BOOK EXCERPT:

A practical handbook packed with expert advice on architectural considerations for designing solutions using SAP BTP to drive digital innovation Purchase of the print or Kindle book includes a free eBook in the PDF format Key FeaturesGuide your customers with proven architectural strategies and considerations on SAP BTPTackle challenges in building process and data integration across complex and hybrid landscapesDiscover SAP BTP services, including visualizations, practical business scenarios, and moreBook Description SAP BTP is the foundation of SAP's intelligent and sustainable enterprise vision for its customers. It's efficient, agile, and an enabler of innovation. It's technically robust, yet its superpower is its business centricity. If you're involved in building IT and business strategies, it's essential to familiarize yourself with SAP BTP to see the big picture for digitalization with SAP solutions. Similarly, if you have design responsibilities for enterprise solutions, learning SAP BTP is crucial to produce effective and complete architecture designs. This book teaches you about SAP BTP in five parts. First, you'll see how SAP BTP is positioned in the intelligent enterprise. In the second part, you'll learn the foundational elements of SAP BTP and find out how it operates. The next part covers integration architecture guidelines, integration strategy considerations, and integration styles with SAP's integration technologies. Later, you'll learn how to use application development capabilities to extend enterprise solutions for innovation and agility. This part also includes digital experience and process automation capabilities. The last part covers how SAP BTP can facilitate data-to-value use cases to produce actionable business insights. By the end of this SAP book, you'll be able to architect solutions using SAP BTP to deliver high business value. What you will learnExplore value propositions and business processes enabled by SAP's Intelligent and Sustainable EnterpriseUnderstand SAP BTP's foundational elements, such as commercial and account modelsDiscover services that can be part of solution designs to fulfill non-functional requirementsGet to grips with integration and extensibility services for building robust solutionsUnderstand what SAP BTP offers for digital experience and process automationExplore data-to-value services that can help manage data and build analytics use casesWho this book is for This SAP guide is for technical architects, solutions architects, and enterprise architects working with SAP solutions to drive digital transformation and innovation with SAP BTP. Some IT background and an understanding of basic cloud concepts is assumed. Working knowledge of the SAP ecosystem will also be beneficial.

Product Details :

Genre : Computers
Author : Serdar Simsekler
Publisher : Packt Publishing Ltd
Release : 2022-10-28
File : 433 Pages
ISBN-13 : 9781801074674


Learning Advanced Python By Studying Open Source Projects

eBook Download

BOOK EXCERPT:

This book is one of its own kind. It is not an encyclopedia or a hands-on tutorial that traps readers in the tutorial hell. It is a distillation of just one common Python user’s learning experience. The experience is packaged with exceptional teaching techniques, careful dependence unraveling and, most importantly, passion. Learning Advanced Python by Studying Open Source Projects helps readers overcome the difficulty in their day-to-day tasks and seek insights from solutions in famous open source projects. Different from a technical manual, this book mixes the technical knowledge, real-world applications and more theoretical content, providing readers with a practical and engaging approach to learning Python. Throughout this book, readers will learn how to write Python code that is efficient, readable and maintainable, covering key topics such as data structures, algorithms, object-oriented programming and more. The author’s passion for Python shines through in this book, making it an enjoyable and inspiring read for both beginners and experienced programmers.

Product Details :

Genre : Computers
Author : Rongpeng Li
Publisher : CRC Press
Release : 2023-11-15
File : 152 Pages
ISBN-13 : 9781000993004


Machine Learning Pocket Reference

eBook Download

BOOK EXCERPT:

With detailed notes, tables, and examples, this handy reference will help you navigate the basics of structured machine learning. Author Matt Harrison delivers a valuable guide that you can use for additional support during training and as a convenient resource when you dive into your next machine learning project. Ideal for programmers, data scientists, and AI engineers, this book includes an overview of the machine learning process and walks you through classification with structured data. You’ll also learn methods for clustering, predicting a continuous value (regression), and reducing dimensionality, among other topics. This pocket reference includes sections that cover: Classification, using the Titanic dataset Cleaning data and dealing with missing data Exploratory data analysis Common preprocessing steps using sample data Selecting features useful to the model Model selection Metrics and classification evaluation Regression examples using k-nearest neighbor, decision trees, boosting, and more Metrics for regression evaluation Clustering Dimensionality reduction Scikit-learn pipelines

Product Details :

Genre : Computers
Author : Matt Harrison
Publisher : O'Reilly Media
Release : 2019-08-27
File : 321 Pages
ISBN-13 : 9781492047513


Data Science And Analytics With Python

eBook Download

BOOK EXCERPT:

Data Science and Analytics with Python is designed for practitioners in data science and data analytics in both academic and business environments. The aim is to present the reader with the main concepts used in data science using tools developed in Python, such as SciKit-learn, Pandas, Numpy, and others. The use of Python is of particular interest, given its recent popularity in the data science community. The book can be used by seasoned programmers and newcomers alike. The book is organized in a way that individual chapters are sufficiently independent from each other so that the reader is comfortable using the contents as a reference. The book discusses what data science and analytics are, from the point of view of the process and results obtained. Important features of Python are also covered, including a Python primer. The basic elements of machine learning, pattern recognition, and artificial intelligence that underpin the algorithms and implementations used in the rest of the book also appear in the first part of the book. Regression analysis using Python, clustering techniques, and classification algorithms are covered in the second part of the book. Hierarchical clustering, decision trees, and ensemble techniques are also explored, along with dimensionality reduction techniques and recommendation systems. The support vector machine algorithm and the Kernel trick are discussed in the last part of the book. About the Author Dr. Jesús Rogel-Salazar is a Lead Data scientist with experience in the field working for companies such as AKQA, IBM Data Science Studio, Dow Jones and others. He is a visiting researcher at the Department of Physics at Imperial College London, UK and a member of the School of Physics, Astronomy and Mathematics at the University of Hertfordshire, UK, He obtained his doctorate in physics at Imperial College London for work on quantum atom optics and ultra-cold matter. He has held a position as senior lecturer in mathematics as well as a consultant in the financial industry since 2006. He is the author of the book Essential Matlab and Octave, also published by CRC Press. His interests include mathematical modelling, data science, and optimization in a wide range of applications including optics, quantum mechanics, data journalism, and finance.

Product Details :

Genre : Computers
Author : Jesus Rogel-Salazar
Publisher : CRC Press
Release : 2018-02-05
File : 308 Pages
ISBN-13 : 9781351647717


Classical Feedback Control

eBook Download

BOOK EXCERPT:

This text describes the design and implementation of high-performance feedback controllers for engineering systems. It emphasizes the frequency-domain design and methods based on Bode integrals, loop shaping and nonlinear dynamic compensation. The book also supplies numerous problems with practcal applications, illustrations and plots, together with MATLAB simulation and design examples.

Product Details :

Genre : Technology & Engineering
Author : Boris Lurie
Publisher : CRC Press
Release : 2000-02-09
File : 480 Pages
ISBN-13 : 0824703707