pynguin
scikit-learn
Our great sponsors
pynguin | scikit-learn | |
---|---|---|
11 | 81 | |
1,197 | 58,130 | |
1.3% | 1.1% | |
8.2 | 9.9 | |
3 days ago | about 21 hours ago | |
Python | Python | |
MIT License | BSD 3-clause "New" or "Revised" License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
pynguin
-
There is framework for everything.
https://swagger.io/specification/ https://github.com/se2p/pynguin
-
Supposed to create tests for a massive project, how should I go about it?
I would use black to reformat this, then, if you can't refactor/rewrite (which is a lot of work!) I would try automated test generation via something like pynguin or fuzzing. I mean … this is not going to be a reliable solution anyways if the codebase is like that. So I would go in a direction that I find interesting to learn about and that could be helpful for the project. That would be generating tests and doing fuzzing. In the end you should run some linters anyways so that you can justify your results and show that the task is not in the scope of an internship and needs extensive refactoring.
-
Klara: Python automatic test generations and static analysis library
The main difference that Klara bring to the table, compared to similar tool like pynguin and Crosshair is that the analysis is entirely static, meaning that no user code will be executed, and you can easily extend the test generation strategy via plugin loading (e.g. the options arg to the Component object returned from function above is not needed for test coverage).
-
Does anybody know a simple algorithm for generating unit tests given a function's code?
Automated White-box test generation software: * https://github.com/EMResearch/EvoMaster -- for integration tests. * https://github.com/se2p/pynguin, https://pynguin.readthedocs.io/en/latest/user/quickstart.html -- unit test generation for python
- se2p/pynguin Pynguin, the PYthoN General UnIt test geNerator, is a tool that allows developers to generate unit tests automatically.
-
Hacker News top posts: Jun 1, 2021
Pynguin – Generate Python unit tests automatically\ (60 comments)
- Pynguin – Generate Python unit tests automatically
- Pynguin – Allow developers to generate Python unit tests automatically
scikit-learn
-
AutoCodeRover resolves 22% of real-world GitHub in SWE-bench lite
Thank you for your interest. There are some interesting examples in the SWE-bench-lite benchmark which are resolved by AutoCodeRover:
- From sympy: https://github.com/sympy/sympy/issues/13643. AutoCodeRover's patch for it: https://github.com/nus-apr/auto-code-rover/blob/main/results...
- Another one from scikit-learn: https://github.com/scikit-learn/scikit-learn/issues/13070. AutoCodeRover's patch (https://github.com/nus-apr/auto-code-rover/blob/main/results...) modified a few lines below (compared to the developer patch) and wrote a different comment.
There are more examples in the results directory (https://github.com/nus-apr/auto-code-rover/tree/main/results).
-
Polars
sklearn is adding support through the dataframe interchange protocol (https://github.com/scikit-learn/scikit-learn/issues/25896). scipy, as far as I know, doesn't explicitly support dataframes (it just happens to work when you wrap a Series in `np.array` or `np.asarray`). I don't know about PyTorch but in general you can convert to numpy.
-
[D] Major bug in Scikit-Learn's implementation of F-1 score
Wow, from the upvotes on this comment, it really seems like a lot of people think that this is the correct behavior! I have to say I disagree, but if that's what you think, don't just sit there upvoting comments on Reddit; instead go to this PR and tell the Scikit-Learn maintainers not to "fix" this "bug", which they are currently planning to do!
- Contraction Clustering (RASTER): A fast clustering algorithm
-
Ask HN: Learning new coding patterns – how to start?
I was in a similar boat to yours - Worked in data science and since then have made a move to data engineering and software engineering for ML services.
I would recommend you look into the Design Patterns book by the Gang of Four. I found it particularly helpful to make extensible code that doesn't break specially with abstract classes, builders and factories. I would also recommend looking into the book The Object Oriented Thought Process to understand why traditional OOP is build the way it is.
You can also look into the source code of popular data science libraries such as sklearn (https://github.com/scikit-learn/scikit-learn/tree/main/sklea...) and see how a lot of them have Base classes to define shared functionality between object of the same nature.
As others mentioned, I would also encourage you to try and implement design patterns in your everyday work - maybe you can make a Factory to load models or preprocessors that follow the same Abstract class?
-
Transformers as Support Vector Machines
It looks like you've been the victim of some misinformation. As Dr_Birdbrain said, an SVM is a convex problem with unique global optimum. sklearn.SVC relies on libsvm which initializes the weights to 0 [0]. The random state is only used to shuffle the data to make probability estimates with Platt scaling [1]. Of the random_state parameter, the sklearn documentation for SVC [2] says
Controls the pseudo random number generation for shuffling the data for probability estimates. Ignored when probability is False. Pass an int for reproducible output across multiple function calls. See Glossary.
[0] https://github.com/scikit-learn/scikit-learn/blob/2a2772a87b...
[1] https://en.wikipedia.org/wiki/Platt_scaling
[2] https://scikit-learn.org/stable/modules/generated/sklearn.sv...
-
How to Build and Deploy a Machine Learning model using Docker
Scikit-learn Documentation
- Planning to get a laptop for ML/DL, is this good enough at the price point or are there better options at/below this price point?
-
Link Prediction With node2vec in Physics Collaboration Network
Firstly, we need a connection to Memgraph so we can get edges, split them into two parts (train set and test set). For edge splitting, we will use scikit-learn. In order to make a connection towards Memgraph, we will use gqlalchemy.
-
WiFilter is a RaspAP install extended with a squidGuard proxy to filter adult content. Great solution for a family, schools and/or public access point
The ML component is based on scikit-learn which differentiates it from purely list-based filters. It couples this with a full-featured wireless router (RaspAP) in a single device, so it fulfills the needs of a use case not entirely addressed by Pi-hole.
What are some alternatives?
CrossHair - An analysis tool for Python that blurs the line between testing and type systems.
Prophet - Tool for producing high quality forecasts for time series data that has multiple seasonality with linear or non-linear growth.
EvoMaster - The first open-source AI-driven tool for automatically generating system-level test cases (also known as fuzzing) for web/enterprise applications. Currently targeting whitebox and blackbox testing of Web APIs, like REST, GraphQL and RPC (e.g., gRPC and Thrift).
Surprise - A Python scikit for building and analyzing recommender systems
klara - Automatic test case generation for python and static analysis library
Keras - Deep Learning for humans
icontract-hypothesis - Combine contracts and automatic testing.
tensorflow - An Open Source Machine Learning Framework for Everyone
methods2test - methods2test is a supervised dataset consisting of Test Cases and their corresponding Focal Methods from a set of Java software repositories
gensim - Topic Modelling for Humans
code - Example application code for the python architecture book
H2O - H2O is an Open Source, Distributed, Fast & Scalable Machine Learning Platform: Deep Learning, Gradient Boosting (GBM) & XGBoost, Random Forest, Generalized Linear Modeling (GLM with Elastic Net), K-Means, PCA, Generalized Additive Models (GAM), RuleFit, Support Vector Machine (SVM), Stacked Ensembles, Automatic Machine Learning (AutoML), etc.