MLRelated.com

SDA Ethernet encapsulation with chatGPT o1 (part 2)

Leonard DieguezLeonard Dieguez March 21, 2025

This is a continuation of the first post on SDA Ethernet encapsulation. In this post we explore reducing the complexity of the module produced and using chatGPT o1 with reasoning. Results are better but are still lacking. The modules still need correction and analysis at a system level for the module.


SDA Ethernet encapsulation with ChatGPT o4

Leonard DieguezLeonard Dieguez February 26, 2025

In this post I explore some of the coding shortcomings of ChatGPT o4. I try to have it code a VHDL module for Ethernet encapsulation based on it reading the SDA specification. As I point out some of the failings it tries to correct the code but also introduces more errors.


Thirty-year coding veteran vs AI

Leonard DieguezLeonard Dieguez October 13, 20231 comment

A thirty-year coding veteran experiments with ChatGPT (GPT-3 and GPT-4) to help implement a VHDL RGMII-to-GMII converter and a 4x oversampling scheme. The post recounts iterative Q&A with the model, shows example VHDL snippets GPT produced, and highlights concrete mistakes the AI made — most notably failing to implement DDR semantics correctly, providing a broken majority-vote function, and overlooking timing/clock-domain nuances. The author concludes that AI can produce useful scaffolding and debugging clues but repeatedly requires careful inspection, simulation, and domain knowledge to correct subtle protocol and timing errors before trusting generated hardware code for real designs.


Why use Area under the curve? (AUC - ROC)

Leonard DieguezLeonard Dieguez September 12, 2023

In scenarios with imbalanced datasets, ROC curves and AUC-ROC scores are valuable tools for assessing and comparing the performance of machine learning classifiers. They help provide insights into a model's ability to distinguish between classes and can guide decision-making regarding threshold selection.


Machine Learning Models Basic Performance Metrics

Leonard DieguezLeonard Dieguez June 10, 2023

When analyzing data using ML, a suitable model is selected based on the task. Classifier models learn from labeled training data and predict discrete classes, while regression models learn from training data and predict continuous values. To evaluate the performance of machine learning models, various metrics are used. These include accuracy, precision, recall, F1 score, AUC-ROC, MAE, MSE, and R-squared. The choice of metrics depends on the specific problem and the nature of the data. Visualization tools such as confusion matrices, ROC curves, precision-recall curves and others can be used to gain insights into the performance of classifiers and understand their behavior. When dealing with imbalanced data, using accuracy as an evaluation metric can be misleading. Accuracy does not account for class imbalance, it may overestimate the performance. It is important to consider other metrics such as AUC and others which provide a more comprehensive evaluation performance in imbalanced datasets.


Thirty-year coding veteran vs AI

Leonard DieguezLeonard Dieguez October 13, 20231 comment

A thirty-year coding veteran experiments with ChatGPT (GPT-3 and GPT-4) to help implement a VHDL RGMII-to-GMII converter and a 4x oversampling scheme. The post recounts iterative Q&A with the model, shows example VHDL snippets GPT produced, and highlights concrete mistakes the AI made — most notably failing to implement DDR semantics correctly, providing a broken majority-vote function, and overlooking timing/clock-domain nuances. The author concludes that AI can produce useful scaffolding and debugging clues but repeatedly requires careful inspection, simulation, and domain knowledge to correct subtle protocol and timing errors before trusting generated hardware code for real designs.


Machine Learning Models Basic Performance Metrics

Leonard DieguezLeonard Dieguez June 10, 2023

When analyzing data using ML, a suitable model is selected based on the task. Classifier models learn from labeled training data and predict discrete classes, while regression models learn from training data and predict continuous values. To evaluate the performance of machine learning models, various metrics are used. These include accuracy, precision, recall, F1 score, AUC-ROC, MAE, MSE, and R-squared. The choice of metrics depends on the specific problem and the nature of the data. Visualization tools such as confusion matrices, ROC curves, precision-recall curves and others can be used to gain insights into the performance of classifiers and understand their behavior. When dealing with imbalanced data, using accuracy as an evaluation metric can be misleading. Accuracy does not account for class imbalance, it may overestimate the performance. It is important to consider other metrics such as AUC and others which provide a more comprehensive evaluation performance in imbalanced datasets.


Why use Area under the curve? (AUC - ROC)

Leonard DieguezLeonard Dieguez September 12, 2023

In scenarios with imbalanced datasets, ROC curves and AUC-ROC scores are valuable tools for assessing and comparing the performance of machine learning classifiers. They help provide insights into a model's ability to distinguish between classes and can guide decision-making regarding threshold selection.


SDA Ethernet encapsulation with chatGPT o1 (part 2)

Leonard DieguezLeonard Dieguez March 21, 2025

This is a continuation of the first post on SDA Ethernet encapsulation. In this post we explore reducing the complexity of the module produced and using chatGPT o1 with reasoning. Results are better but are still lacking. The modules still need correction and analysis at a system level for the module.


SDA Ethernet encapsulation with ChatGPT o4

Leonard DieguezLeonard Dieguez February 26, 2025

In this post I explore some of the coding shortcomings of ChatGPT o4. I try to have it code a VHDL module for Ethernet encapsulation based on it reading the SDA specification. As I point out some of the failings it tries to correct the code but also introduces more errors.