|
--- |
|
base_model: sentence-transformers/paraphrase-mpnet-base-v2 |
|
library_name: setfit |
|
metrics: |
|
- f1 |
|
pipeline_tag: text-classification |
|
tags: |
|
- setfit |
|
- sentence-transformers |
|
- text-classification |
|
- generated_from_setfit_trainer |
|
widget: |
|
- text: 'Title: "Deep Residual Learning for Image Recognition". |
|
|
|
Abstract: In this paper, we propose a new deep residual learning framework for |
|
image classification. We introduce a novel residual block architecture that learns |
|
to represent high-level features in an image. Our approach is based on the idea |
|
of residual learning, where the network learns to represent the difference between |
|
the input and the output of a layer, rather than learning to represent the output |
|
directly. We evaluate our approach on several benchmark datasets, including ImageNet |
|
and CIFAR-10, and show that it achieves state-of-the-art performance. Our results |
|
demonstrate the effectiveness of residual learning for image classification, and |
|
show that it can be used to improve the performance of deep neural networks. We |
|
also provide a detailed analysis of the residual block architecture, and show |
|
how it can be used to improve the performance of other deep learning models. This |
|
paper provides a comprehensive overview of the residual learning framework, and |
|
demonstrates its effectiveness for image classification tasks.' |
|
- text: Let G be a finite group and let V be a finite-dimensional representation of |
|
G over an algebraically closed field k. We say that V is a representation of G |
|
in characteristic zero if the characteristic of k is zero. In this paper, we investigate |
|
the structure of the representation ring R(G) of a finite group G in characteristic |
|
zero. We show that R(G) is isomorphic to the group ring k[G] if and only if G |
|
is a cyclic group. Furthermore, we provide a characterization of the representation |
|
rings of finite abelian groups in terms of their irreducible representations. |
|
Our results have implications for the study of the representation theory of finite |
|
groups in characteristic zero. |
|
- text: 'Denotational Semantics of Programming Languages: A Survey |
|
|
|
|
|
Abstract: Denotational semantics is a branch of programming language theory that |
|
focuses on the meaning of programming languages. In this survey, we provide an |
|
overview of the key concepts and results in denotational semantics, including |
|
the use of domain theory and categorical semantics. We also discuss the relationship |
|
between denotational semantics and other areas of programming language theory, |
|
such as operational semantics and axiomatic semantics. |
|
|
|
|
|
Introduction |
|
|
|
|
|
Denotational semantics is a mathematical approach to understanding the meaning |
|
of programming languages. It is based on the idea that a programming language |
|
can be viewed as a mapping from a set of programs to a set of meanings, where |
|
the meanings are represented as mathematical objects. The key concept in denotational |
|
semantics is the notion of a denotation, which is a function that maps a program |
|
to its meaning. |
|
|
|
|
|
Domain Theory |
|
|
|
|
|
Domain theory is a branch of mathematics that provides a framework for understanding |
|
the notion of a denotation. It is based on the idea that a denotation is a function |
|
from a set of programs to a set of meanings, where the meanings are represented |
|
as elements of a domain. The key concept in domain theory is the notion of a continuous |
|
function, which is a function that preserves the order relation on the domain. |
|
|
|
|
|
Categorical Semantics |
|
|
|
|
|
Categorical semantics is a branch of mathematics that provides a framework for |
|
understanding the notion of a denotation in a categorical setting. It is based |
|
on the idea that a denotation is a function from a set of programs to a set of |
|
meanings, where the meanings are represented as objects in a category. The key |
|
concept in categorical semantics is the notion of a functor, which is a function |
|
that preserves the morphisms in the category. |
|
|
|
|
|
Conclusion |
|
|
|
|
|
In this survey, we have provided an overview of the key concepts and results in |
|
denotational semantics. We have also discussed the relationship between denotational |
|
semantics and other areas of programming language theory. The results presented |
|
in this survey demonstrate the importance of denotational semantics in understanding |
|
the meaning of programming languages.' |
|
- text: 'A Novel Robust Control Approach for Uncertain Systems with Time-Varying Delays |
|
|
|
|
|
Abstract: This paper presents a new robust control method for uncertain systems |
|
with time-varying delays. The proposed approach combines the advantages of model |
|
predictive control (MPC) and sliding mode control (SMC) to achieve robust stability |
|
and performance. The MPC algorithm is used to predict the future behavior of the |
|
system, while the SMC algorithm is employed to reject disturbances and uncertainties. |
|
The stability and performance of the proposed controller are analyzed using Lyapunov |
|
theory and simulation results. The effectiveness of the proposed approach is demonstrated |
|
through numerical examples and comparisons with existing methods. |
|
|
|
|
|
Keywords: Robust control, uncertain systems, time-varying delays, model predictive |
|
control, sliding mode control, Lyapunov theory.' |
|
- text: 'A Novel Compiler Framework for Parallel Computing: Design and Implementation |
|
|
|
|
|
Abstract: With the increasing demand for high-performance computing, parallel |
|
computing has become a crucial aspect of modern computing systems. However, the |
|
complexity of parallel programming models and the lack of efficient compilation |
|
techniques hinder the widespread adoption of parallel computing. In this paper, |
|
we propose a novel compiler framework for parallel computing, which aims to bridge |
|
the gap between parallel programming models and efficient compilation techniques. |
|
Our framework, called ParComp, is designed to support a wide range of parallel |
|
programming models, including OpenMP, MPI, and CUDA. ParComp consists of three |
|
main components: a parallelization module, a scheduling module, and a code generation |
|
module. The parallelization module is responsible for identifying parallelizable |
|
loops and transforming them into parallel code. The scheduling module is responsible |
|
for scheduling the parallel tasks and allocating resources to them. The code generation |
|
module is responsible for generating efficient parallel code from the scheduled |
|
tasks. We evaluate the performance of ParComp using a set of benchmark programs |
|
and compare it with state-of-the-art parallel compilers. The results show that |
|
ParComp outperforms the existing compilers in terms of execution time and scalability. |
|
Our framework is implemented using C++ and is available online for public use. |
|
|
|
|
|
Index Terms: Parallel computing, compiler design, parallel programming models, |
|
OpenMP, MPI, CUDA, code generation, scheduling, resource allocation.' |
|
inference: true |
|
model-index: |
|
- name: SetFit with sentence-transformers/paraphrase-mpnet-base-v2 |
|
results: |
|
- task: |
|
type: text-classification |
|
name: Text Classification |
|
dataset: |
|
name: Unknown |
|
type: unknown |
|
split: test |
|
metrics: |
|
- type: f1 |
|
value: 0.6184 |
|
name: F1 |
|
--- |
|
|
|
# SetFit with sentence-transformers/paraphrase-mpnet-base-v2 |
|
|
|
This is a [SetFit](https://github.com/huggingface/setfit) model that can be used for Text Classification. This SetFit model uses [sentence-transformers/paraphrase-mpnet-base-v2](https://huggingface.co/sentence-transformers/paraphrase-mpnet-base-v2) as the Sentence Transformer embedding model. A [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance is used for classification. |
|
|
|
The model has been trained using an efficient few-shot learning technique that involves: |
|
|
|
1. Fine-tuning a [Sentence Transformer](https://www.sbert.net) with contrastive learning. |
|
2. Training a classification head with features from the fine-tuned Sentence Transformer. |
|
|
|
## Model Details |
|
|
|
### Model Description |
|
- **Model Type:** SetFit |
|
- **Sentence Transformer body:** [sentence-transformers/paraphrase-mpnet-base-v2](https://huggingface.co/sentence-transformers/paraphrase-mpnet-base-v2) |
|
- **Classification head:** a [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance |
|
- **Maximum Sequence Length:** 512 tokens |
|
- **Number of Classes:** 11 classes |
|
<!-- - **Training Dataset:** [Unknown](https://huggingface.co/datasets/unknown) --> |
|
<!-- - **Language:** Unknown --> |
|
<!-- - **License:** Unknown --> |
|
|
|
### Model Sources |
|
|
|
- **Repository:** [SetFit on GitHub](https://github.com/huggingface/setfit) |
|
- **Paper:** [Efficient Few-Shot Learning Without Prompts](https://arxiv.org/abs/2209.11055) |
|
- **Blogpost:** [SetFit: Efficient Few-Shot Learning Without Prompts](https://huggingface.co/blog/setfit) |
|
|
|
### Model Labels |
|
| Label | Examples | |
|
|:--------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------| |
|
| Data Structures | <ul><li>'Title: On the Complexity of Gröbner Bases for Toric Ideals\n\nAbstract: We investigate the computational complexity of computing Gröbner bases for toric ideals. Our main result is a polynomial-time algorithm for computing Gröbner bases for toric ideals in the case where the toric ideal is generated by a set of binomials. We also show that this algorithm can be used to solve a number of problems in computational algebra, including the computation of the Hilbert series of a toric ideal and the determination of the dimension of a toric variety. Our results have implications for the study of toric varieties and their applications in computer science and engineering.\n\nIntroduction\n\nToric varieties are a fundamental object of study in algebraic geometry, and have found numerous applications in computer science and engineering. In this paper, we investigate the computational complexity of computing Gröbner bases for toric ideals, which are a key tool in the study of toric varieties. Our main result is a polynomial-time algorithm for computing Gröbner bases for toric ideals in the case where the toric ideal is generated by a set of binomials.\n\nBackground\n\nA toric ideal is a polynomial ideal that is generated by a set of binomials. The Gröbner basis of a toric ideal is a set of binomials that generate the ideal and have a certain property called the S-polynomial property. The S-polynomial property is a key tool in the study of toric varieties, and has been used to solve a number of problems in computational algebra.\n\nMain Result\n\nOur main result is a polynomial-time algorithm for computing Gröbner bases for toric ideals in the case where the toric ideal is generated by a set of binomials. The algorithm works by first computing the Hilbert series of the toric ideal, and then using this information to compute the Gröbner basis. We show that this algorithm can be used to solve a number of problems in computational algebra, including the computation of the Hilbert series of a toric ideal and the determination of the dimension of a toric variety.\n\nConclusion\n\nIn this paper, we have investigated the computational complexity of computing Gröbner bases for toric ideals. Our main result is a polynomial-time algorithm for computing Gröbner bases for toric ideals in the case where the toric ideal is generated by a set of binomials. We believe that this result has implications for the study of toric varieties and their applications in computer science and engineering.'</li><li>"A novel approach to designing efficient hash tables for large-scale data storage is proposed in this paper. The proposed hash table, dubbed 'Efficient Hash Table' (EHT), employs a combination of open addressing and linear probing to minimize collisions and improve search times. Experimental results demonstrate that EHT outperforms existing hash table implementations in terms of search time and memory usage, making it an attractive solution for big data applications. The EHT algorithm is implemented using a C++ programming language and is shown to scale well on multi-core processors. This paper contributes to the field of data structures by providing a new, efficient, and scalable hash table design that can be used in a variety of applications, including databases, file systems, and cloud storage systems."</li><li>"Abstract: This paper presents an efficient array-based algorithm for searching and sorting large datasets. The proposed algorithm utilizes a combination of bit-packing and prefix sums to achieve a time complexity of O(n) for search operations and O(n log n) for sort operations. Experimental results demonstrate the algorithm's superiority over existing methods in terms of performance and memory usage. The algorithm is particularly useful for applications where data is stored in arrays, such as in computer vision and scientific simulations. The proposed algorithm is implemented in C++ and is available for download from the IEEE Xplore digital library."</li></ul> | |
|
| Programming Languages | <ul><li>'Title: A Functional Programming Approach to Type Inference in Higher-Order Logic Programming Languages.\nAbstract: This paper presents a novel approach to type inference in higher-order logic programming languages using functional programming techniques. We propose a type system that combines the benefits of higher-order logic programming with the expressive power of functional programming. Our approach is based on a novel type inference algorithm that uses a combination of type reconstruction and type checking. We demonstrate the effectiveness of our approach through a series of experiments on a set of benchmark programs. The results show that our approach outperforms existing type inference systems in terms of accuracy and efficiency. We also discuss the implications of our work for the design of future programming languages.'</li><li>'A Comparative Study of Functional Programming Paradigms in Haskell and Scala for Efficient Software Development\nAbstract: Functional programming has gained significant attention in recent years due to its ability to promote modular, composable, and reusable code. This paper presents a comparative study of two popular functional programming languages, Haskell and Scala, with a focus on their application in efficient software development. We analyze the strengths and weaknesses of each language, highlighting their respective features and performance characteristics. Our results demonstrate that Haskell and Scala can be effectively used for building high-performance software systems, with Haskell exhibiting superior performance in certain scenarios. The findings of this study contribute to the ongoing debate on the choice of functional programming languages for software development and provide insights for practitioners and researchers alike.\nKeywords: functional programming, Haskell, Scala, software development, performance evaluation\n'</li><li>'A Novel Type Theory for Dependent Type Systems in Programming Languages\n\nAbstract: This paper proposes a novel type theory for dependent type systems, which is a fundamental component of programming languages. The proposed type theory is based on a combination of ideas from homotopy type theory and dependent type theory. We show that the proposed type theory is sound and complete, and we provide a formal proof of its soundness. We also demonstrate the expressiveness of the proposed type theory by implementing a dependent type checker using it. The results of this paper demonstrate the potential of the proposed type theory for use in programming languages.\n\nKeywords: dependent type systems, programming languages, type theory, homotopy type theory, dependent type theory.\n\n1 Introduction\n\nDependent type systems are a fundamental component of programming languages, allowing programmers to specify and reason about the types of complex data structures. However, the design of dependent type systems is challenging due to the need to balance expressiveness and decidability. In this paper, we propose a novel type theory for dependent type systems, which is based on a combination of ideas from homotopy type theory and dependent type theory.\n\n2 Background\n\nDependent type systems are based on the concept of dependent types, which are types that depend on the values of other types. Dependent types are used to specify the types of complex data structures, such as lists and matrices. However, the design of dependent type systems is challenging due to the need to balance expressiveness and decidability.\n\n3 Proposed Type Theory\n\nIn this paper, we propose a novel type theory for dependent type systems, which is based on a combination of ideas from homotopy type theory and dependent type theory. The proposed type theory is sound and complete, and we provide a formal proof of its soundness. We also demonstrate the expressiveness of the proposed type theory by implementing a dependent type checker using it.\n\n4 Conclusion\n\nThe results of this paper demonstrate the potential of the proposed type theory for use in programming languages. The proposed type theory is sound and complete, and it provides a formal foundation for dependent type systems. We believe that the proposed type theory has the potential to be used in a wide range of programming languages, and we plan to continue exploring its applications in the future.\n\nReferences:\n\n[1] N. Ghani and P. J. Scott, "A type theory for dependent types," in Proceedings of the 22nd Annual Symposium on Logic in Computer Science, 2007, pp. 233-242.\n[2] P. J. Scott, "A type theory for dependent types," Ph.D. dissertation, University of Edinburgh, 2007.\n[3] A. K. Ghosh and P. J. Scott, "A type theory for dependent types," Journal of Functional Programming, vol. 19, no. 3-4, pp. 437-462, 2009.\n[4] P. J. Scott, "A type theory for dependent types," in Proceedings of the 25th Annual Symposium on Logic in Computer Science, 2010, pp. 231-240.\n[5] A. K. Ghosh and P. J. Scott, "A type theory for dependent types," Journal of Functional Programming, vol. 22, no. 2-3, pp. 147-172, 2012.\n[6] P. J. Scott, "A type theory for dependent types," in Proceedings of the 28th Annual Symposium on Logic in Computer Science, 2013, pp. 231-240.\n[7] A. K. Ghosh and P. J. Scott, "A type theory for dependent types," Journal of Functional Programming, vol. 25, no. 2-3, pp. 147-172, 2015.\n[8] P. J. Scott, "A type theory for dependent types," in Proceedings of the 31st Annual Symposium on Logic in Computer Science, 2016, pp. 231-240.\n[9] A. K. Ghosh and P. J. Scott, "A type theory for dependent types," Journal of Functional Programming, vol. 28, no. 2-3, pp. 147-172, 2018.\n[10] P. J. Scott, "A type theory for dependent types," in Proceedings of the 34th Annual Symposium on Logic in Computer Science, 2019, pp. 231-240.\n\n'</li></ul> | |
|
| Information Theory | <ul><li>'A Novel Approach to Secure Data Transmission Using Quantum Key Distribution\n\nAbstract: This paper proposes a novel approach to secure data transmission using quantum key distribution (QKD). The proposed scheme utilizes the principles of quantum mechanics to enable secure key exchange between two parties. The scheme is based on the BB84 protocol, which is a widely used QKD protocol. However, the proposed scheme introduces a new feature, known as the "quantum error correction" mechanism, which enables the detection of any eavesdropping attempts. The proposed scheme is analyzed using the security analysis framework, which shows that it is secure against any eavesdropping attempts. The performance of the proposed scheme is evaluated using simulations, which show that it outperforms existing QKD schemes in terms of key generation rate and security. The proposed scheme has the potential to be used in various applications, including secure data transmission over the internet.'</li><li>'A Secure Communication Framework for IoT Devices using Quantum Key Distribution and Homomorphic Encryption\n\nAbstract: The Internet of Things (IoT) has revolutionized the way we live and work, connecting billions of devices worldwide. However, the increased connectivity also raises significant security concerns, as IoT devices are often vulnerable to cyber attacks. In this paper, we propose a secure communication framework for IoT devices using quantum key distribution (QKD) and homomorphic encryption. Our framework leverages the principles of QKD to establish secure key exchange between IoT devices, while homomorphic encryption enables secure data processing without revealing sensitive information. We demonstrate the effectiveness of our framework through simulations and experiments, showcasing its ability to provide robust security against various types of attacks. The results of this study contribute to the development of secure IoT communication systems, which are essential for the widespread adoption of IoT technology.\n\nKeywords: Quantum Key Distribution, Homomorphic Encryption, IoT Security, Secure Communication Framework\n\nI. Introduction\n\nThe Internet of Things (IoT) has transformed the way we live and work, connecting billions of devices worldwide. However, the increased connectivity also raises significant security concerns, as IoT devices are often vulnerable to cyber attacks. In this paper, we propose a secure communication framework for IoT devices using quantum key distribution (QKD) and homomorphic encryption.\n\nII. Background\n\nQuantum Key Distribution (QKD) is a method of secure key exchange that relies on the principles of quantum mechanics. QKD enables two parties to establish a shared secret key over an insecure communication channel, without revealing the key to any third party. Homomorphic encryption, on the other hand, is a type of encryption that enables computations to be performed on encrypted data without decrypting it first.\n\nIII. Proposed Framework\n\nOur proposed framework consists of two main components: QKD-based key exchange and homomorphic encryption-based data processing. The QKD component uses the BB84 protocol to establish a secure key exchange between IoT devices, while the homomorphic encryption component uses the HElib library to perform secure data processing.\n\nIV. Simulation and Experiment Results\n\nWe conducted simulations and experiments to evaluate the effectiveness of our framework. The results show that our framework provides robust security against various types of attacks, including eavesdropping and tampering attacks. We also demonstrate the ability of our framework to provide secure data processing without revealing sensitive information.\n\nV. Conclusion\n\nIn this paper, we proposed a secure communication framework for IoT devices using QKD and homomorphic encryption. Our framework provides robust security against various types of attacks and enables secure data processing without revealing sensitive information. The results of this study contribute to the development of secure IoT communication systems, which are essential for the widespread adoption of IoT technology.'</li><li>'A Novel Turbo Code Design for Near-Capacity Performance in Wireless Communication Systems\nAbstract—Turbo codes have been widely adopted in various wireless communication systems due to their near-capacity performance and low complexity. In this paper, we propose a novel turbo code design that achieves better performance than the traditional turbo code. The proposed design is based on a new interleaving scheme that combines the benefits of random and systematic interleaving. Simulation results show that the proposed turbo code outperforms the traditional turbo code in terms of bit error rate and frame error rate. The proposed design is also compared with other state-of-the-art turbo code designs, and the results show that it achieves better performance. The proposed turbo code is suitable for various wireless communication systems, including 5G and beyond. The design and implementation of the proposed turbo code are discussed in detail, and the simulation results are presented to demonstrate its performance.\nKeywords—Turbo codes, interleaving scheme, near-capacity performance, wireless communication systems, 5G and beyond.'</li></ul> | |
|
| Group Theory | <ul><li>'Title: On the Structure of the Centralizer of a Toral Subgroup in a Reductive Algebraic Group\n\nAbstract: We study the centralizer of a toral subgroup in a reductive algebraic group over an algebraically closed field of characteristic zero. Our main result describes the structure of this centralizer in terms of the root system of the group and the weights of the toral subgroup. We also provide a characterization of the centralizer in terms of the Bruhat-Tits building of the group. Our methods involve a combination of Lie algebra techniques, including the use of the Killing form and the Cartan-Killing classification of simple Lie algebras, as well as geometric and algebraic techniques, including the use of the Bruhat-Tits building and the theory of algebraic groups over local fields.\n\nIntroduction: The centralizer of a toral subgroup in a reductive algebraic group is a fundamental object of study in the theory of algebraic groups. In this paper, we investigate the structure of this centralizer, with a focus on the case where the toral subgroup is a maximal torus. Our main result provides a detailed description of the centralizer in this case, and we also provide a characterization of the centralizer in terms of the Bruhat-Tits building of the group. We hope that our results will be of interest to researchers in the field of algebraic groups and Lie theory.\n\n1 Introduction\n\n1.1 Background and Motivation\n\n1.2 Main Results\n\n2 Preliminaries\n\n2.1 Algebraic Groups and Lie Algebras\n\n2.2 Root Systems and Weights\n\n2.3 Bruhat-Tits Buildings\n\n3 The Centralizer of a Toral Subgroup\n\n3.1 Definition and Basic Properties\n\n3.2 Structure of the Centralizer\n\n3.3 Characterization of the Centralizer\n\n4 Applications and Further Directions\n\n4.1 Applications to Algebraic Groups\n\n4.2 Further Directions\n\nReferences\n\nBibliography'</li><li>'Geometric Invariant Theory (GIT) is a branch of algebraic geometry that studies the action of an algebraic group on an algebraic variety. It provides a framework for understanding the symmetries of algebraic varieties and has applications in various areas of mathematics, including commutative algebra, algebraic geometry, and representation theory. In this paper, we apply GIT to study the invariant theory of a certain algebraic group action on a projective variety. We use the Hilbert-Mumford criterion to determine the semistable points of the action and then compute the invariant ring using the Grothendieck ring of the group. Our results have implications for the study of algebraic groups and their actions on projective varieties.'</li><li>'Title: A Geometric Approach to Representations of Finite Groups\n\nAbstract: We introduce a new method for constructing representations of finite groups using algebraic geometry. Our approach is based on the idea of representing a group as a quotient of a reductive group by a finite subgroup. We show that this construction yields a faithful representation of the group, and we use it to compute the character table of the symmetric group S_5. Our method has several advantages over existing methods, including the ability to handle large groups and the flexibility to incorporate additional structure. We also discuss some potential applications of our method, including the computation of representation theory for finite groups of Lie type.\n\nKeywords: representation theory, algebraic geometry, finite groups, reductive groups, symmetric group\n\nArXiv ID: 2203.10201\n\nSubmission date: 2022-03-17'</li></ul> | |
|
| Neural and Evolutionary | <ul><li>'Evolutionary algorithms have been widely used in various optimization problems due to their ability to efficiently search for optimal solutions. In this paper, we propose a novel hybrid approach that combines the strengths of genetic algorithms and differential evolution to solve complex optimization problems. The proposed method, called GEDE, integrates the exploration capabilities of genetic algorithms with the exploitation capabilities of differential evolution. We evaluate the performance of GEDE on several benchmark problems and compare it with other state-of-the-art algorithms. The results show that GEDE outperforms the other algorithms in terms of convergence speed and solution quality. We also analyze the convergence behavior of GEDE and provide insights into its performance. The proposed approach has the potential to be applied to a wide range of optimization problems in various fields, including engineering, economics, and computer science.'</li><li>'Title: A Deep Learning Approach for Sentiment Analysis of Text Data\nAbstract: This paper proposes a novel deep learning model for sentiment analysis of text data. The proposed model combines the strengths of convolutional neural networks (CNNs) and recurrent neural networks (RNNs) to effectively capture the spatial and temporal dependencies in text data. Experimental results on several benchmark datasets demonstrate the superiority of the proposed model over state-of-the-art methods. The proposed model achieves an accuracy of 92.5% on the IMDB dataset, outperforming the best existing method by 2.5%. The results also show that the proposed model is robust to noise and can handle out-of-vocabulary words. The proposed model is a significant contribution to the field of natural language processing and has the potential to be applied to various real-world applications.'</li><li>'Title: A Novel Hybrid Approach for Deep Learning-based Optimization of Evolutionary Algorithms\n\nAbstract: This paper proposes a novel hybrid approach that combines the strengths of deep learning and evolutionary algorithms to optimize complex optimization problems. We introduce a new neural network architecture that learns to adapt the parameters of evolutionary algorithms in real-time, leading to improved convergence rates and better solution quality. Our approach is evaluated on a range of benchmark problems and compared to state-of-the-art methods. The results show that our hybrid approach outperforms existing methods in terms of convergence speed and solution quality. We also provide a comprehensive analysis of the proposed approach and discuss its potential applications in various fields.\n\nKeywords: Evolutionary algorithms, Deep learning, Optimization, Hybrid approach, Neural networks.'</li></ul> | |
|
| Commutative Algebra | <ul><li>"Title: On the Frobenius Ideals of a Commutative Ring. Abstract: We study the Frobenius ideals of a commutative ring and provide a characterization of the Frobenius ideals in terms of the ring's structure. Our main result shows that the Frobenius ideal of a commutative ring is a finitely generated ideal if and only if the ring is a finitely generated module over its endomorphism ring. We also investigate the relationship between the Frobenius ideal and the ring's dimension. Our results have implications for the study of commutative algebra and the theory of Frobenius ideals. Keywords: Frobenius ideal, commutative ring, finitely generated ideal, endomorphism ring, dimension. Source: Google Scholar."</li><li>'arXiv:2207.12345v1 [math.AG] 18 Jul 2022\n\nTitle: On the arithmetic of certain K3 surfaces\n\nAbstract: We study the arithmetic of certain K3 surfaces defined over the rationals, with a focus on their Picard groups and Neron-Severi lattices. Our main result is a complete classification of these surfaces in terms of their invariants, which we compute using a combination of algebraic geometry and number theory techniques. We also provide explicit examples of such surfaces, and discuss their implications for the study of arithmetic geometry.\n\nIntroduction\n\nLet $S$ be a K3 surface defined over the rationals, and let $Pic(S)$ denote its Picard group. The Neron-Severi lattice $NS(S)$ is the subgroup of $Pic(S)$ generated by the divisor classes of the curve $S$. Our main goal is to classify the K3 surfaces $S$ such that $NS(S)$ is isomorphic to a lattice of the form $U \times E_8$, where $U$ is a hyperbolic plane and $E_8$ is the standard $E_8$ lattice. We achieve this by first showing that such a surface must have a certain type of singular point, and then using this information to compute the invariants of $NS(S)$. Our classification result has implications for the study of arithmetic geometry, and provides a new perspective on the geometry of K3 surfaces.'</li><li>'arXiv:2203.01023v1 [math.RT] 1 Mar 2022\n\nTitle: On the representation theory of the Iwahori-Hecke algebra of the symmetric group\n\nAbstract: We study the representation theory of the Iwahori-Hecke algebra of the symmetric group. Our main result is a classification of the irreducible representations of this algebra in terms of the representation theory of the symmetric group. We also provide a new proof of the fact that the Iwahori-Hecke algebra is a semisimple algebra. Our methods involve a combination of representation theory, algebraic geometry, and combinatorics.\n\n1 Introduction\n\nThe Iwahori-Hecke algebra of the symmetric group is a well-studied algebra that has connections to many areas of mathematics, including representation theory, algebraic geometry, and combinatorics. In this paper, we study the representation theory of this algebra, with a focus on classifying the irreducible representations. Our main result is a classification of the irreducible representations of the Iwahori-Hecke algebra in terms of the representation theory of the symmetric group. We also provide a new proof of the fact that the Iwahori-Hecke algebra is a semisimple algebra. Our methods involve a combination of representation theory, algebraic geometry, and combinatorics.\n\n2 Background\n\nIn this section, we provide some background on the representation theory of the symmetric group and the Iwahori-Hecke algebra. We recall the definition of the Iwahori-Hecke algebra and its properties, and we also recall some results on the representation theory of the symmetric group.\n\n3 The Representation Theory of the Iwahori-Hecke Algebra\n\nIn this section, we study the representation theory of the Iwahori-Hecke algebra. We provide a classification of the irreducible representations of this algebra in terms of the representation theory of the symmetric group. We also provide a new proof of the fact that the Iwahori-Hecke algebra is a semisimple algebra.\n\n4 Conclusion\n\nIn this paper, we have studied the representation theory of the Iwahori-Hecke algebra of the symmetric group. Our main result is a classification of the irreducible representations of this algebra in terms of the representation theory of the symmetric group. We have also provided a new proof of the fact that the Iwahori-Hecke algebra is a semisimple algebra. We believe that our results will have applications in many areas of mathematics, including representation theory, algebraic geometry, and combinatorics.\n\nReferences\n\n[1] Ariki, S. (1996). On the decomposition numbers of the Hecke algebra of the symmetric group. Journal of Algebra, 183(2), 371-394.\n[2] Ariki, S., & Koike, K. (1994). A Hecke algebra of (Z/rZ)Sn and approximation of the irreducible characters of the symmetric group. Journal of Algebra, 171(2), 311-346.\n[3] Dipper, R., & James, G. (1988). Representations of the symmetric group which are irreducible over the commutator subgroup. Mathematische Zeitschrift, 198(2), 151-166.\n[4] Green, J. A. (1955). Axiomatic approach to the representation theory of the symmetric group. Journal of Algebra, 1(2), 107-133.\n[5] James, G. (1978). The representation theory of the symmetric group. Lecture Notes in Mathematics, 682, 1-42.\n[6] Murphy, G. J. (1990). On the representation theory of the symmetric group. Journal of Algebra, 131(2), 449-465.\n[7] Nakayama, T. (1952). On the representations of the symmetric group. Journal of the Faculty of Science, University of Tokyo, 6(2), 147-172.\n[8] Sagan, B. E. (1991). The symmetric group: Representations, combinatorial algorithms, and symmetric functions. Wadsworth & Brooks/Cole.\n[9] Zelevinsky, A. (1980). Representations of the symmetric group which are irreducible over the commutator subgroup. Mathematische Zeitschrift, 173(2), 133-146.\n[10] Zelevinsky, A. (1981). Representations of the symmetric group which are irreducible over the commutator subgroup. Journal of Algebra, 71(2), 249-262.\n\n\n'</li></ul> | |
|
| Systems and Control | <ul><li>'Optimal Control Theory for Nonlinear Systems with Bounded Controls\n\nAbstract: This paper presents a new approach to optimal control theory for nonlinear systems with bounded controls. We propose a novel method for solving the Hamilton-Jacobi-Isaacs equation, which is a fundamental equation in optimal control theory. Our approach is based on a combination of deep learning and numerical methods, and it is capable of handling high-dimensional systems with nonlinear dynamics. We demonstrate the effectiveness of our method through numerical experiments on several benchmark problems, including a nonlinear pendulum and a nonlinear cart-pole system. Our results show that our method can achieve better performance than existing methods, and it is computationally efficient. We also provide a theoretical analysis of our method, and we show that it converges to the optimal solution under certain conditions. The proposed method has the potential to be applied to a wide range of fields, including robotics, aerospace engineering, and biomedical engineering.'</li><li>'Title: A Robust Control Approach for Uncertain Systems with Time-Varying Delays\nAbstract: This paper presents a robust control strategy for uncertain systems with time-varying delays. The proposed method combines a model predictive control (MPC) scheme with a robust control approach to ensure stability and performance of the closed-loop system. The MPC scheme is designed to handle the time-varying delays, while the robust control approach ensures that the system remains stable despite the presence of uncertainties. The effectiveness of the proposed method is demonstrated through numerical simulations and experimental results on a laboratory setup. The results show that the proposed method outperforms traditional robust control approaches in terms of stability and performance. The proposed method has the potential to be applied to a wide range of uncertain systems with time-varying delays, such as those encountered in robotics, aerospace, and process control.'</li><li>'H-infinity Control in Robust Control\n\nAbstract: This paper presents a novel approach to robust control design using H-infinity control theory. The proposed method combines the advantages of both H-infinity and mu-synthesis techniques to achieve improved robustness and performance. Theoretical results are supported by numerical examples and comparisons with existing methods. The proposed approach is demonstrated on a benchmark problem and shows significant improvements in terms of robust stability and performance.\n\nIntroduction\n\nH-infinity control theory has been widely used in robust control design due to its ability to provide guaranteed robust stability and performance. However, the existing methods often suffer from conservativeness and computational complexity. In this paper, we propose a novel approach that combines the advantages of H-infinity and mu-synthesis techniques to achieve improved robustness and performance. The proposed method is based on a new optimization problem formulation that takes into account the uncertainties and disturbances in the system.\n\nMethodology\n\nThe proposed approach is based on the following steps: (1) model the system using a state-space representation; (2) formulate the H-infinity control problem using the mu-synthesis framework; (3) solve the optimization problem to obtain the controller gains; and (4) implement the controller using a digital signal processor. The proposed approach is demonstrated on a benchmark problem, which is a two-input two-output system with uncertainties in the plant and disturbances in the input.\n\nResults\n\nThe proposed approach is compared with existing methods, including H-infinity control and mu-synthesis. The results show that the proposed approach achieves improved robust stability and performance, with a significant reduction in the control effort. The proposed approach is also compared with a state-of-the-art robust control method, which shows that the proposed approach outperforms the existing method in terms of robust stability and performance.\n\nConclusion\n\nIn this paper, we have presented a novel approach to robust control design using H-infinity control theory. The proposed method combines the advantages of both H-infinity and mu-synthesis techniques to achieve improved robustness and performance. Theoretical results are supported by numerical examples and comparisons with existing methods. The proposed approach is demonstrated on a benchmark problem and shows significant improvements in terms of robust stability and performance.'</li></ul> | |
|
| Statistics Theory | <ul><li>'Title: A New Perspective on the Generalization Error of Support Vector Machines\nAbstract: We provide a new bound on the generalization error of support vector machines (SVMs) in terms of the Rademacher complexity of the reproducing kernel Hilbert space (RKHS) of the kernel. Our bound is tighter than existing bounds and has a simpler form. We also provide a new algorithm for learning the kernel, which is based on the idea of minimizing the empirical risk with respect to the RKHS norm. We demonstrate the effectiveness of our approach on several benchmark datasets.'</li><li>'Title: A Bayesian Approach to Hypothesis Testing for High-Dimensional Data\n\nAbstract: Hypothesis testing is a fundamental problem in statistics, and its applications are widespread in various fields. However, the traditional methods of hypothesis testing often fail to perform well in high-dimensional data settings. In this paper, we propose a novel Bayesian approach to hypothesis testing for high-dimensional data. Our method combines the strengths of Bayesian inference and dimensionality reduction techniques to provide a robust and efficient solution to the hypothesis testing problem. We demonstrate the effectiveness of our approach through extensive simulations and real-world experiments on high-dimensional data sets. The results show that our method outperforms existing methods in terms of accuracy and computational efficiency. Furthermore, we provide a theoretical analysis of our approach, which provides insights into its performance and limitations. Our method has the potential to be applied to a wide range of applications, including image analysis, genomics, and finance. The code and data used in this paper are available online for reproducibility purposes.'</li><li>'Title: Bayesian Network Learning with Gaussian Process Priors for Uncertainty Quantification in High-Dimensional Systems\n\nAbstract: Bayesian networks are a powerful tool for modeling complex systems with uncertainty. However, in high-dimensional systems, the computational cost of learning Bayesian networks can be prohibitively expensive. In this paper, we propose a novel approach to Bayesian network learning using Gaussian process priors. Our approach, which we call Bayesian network learning with Gaussian process priors (BN-GP), leverages the flexibility of Gaussian processes to model the uncertainty in the network structure. We demonstrate the effectiveness of BN-GP on several high-dimensional systems, including a synthetic dataset and a real-world dataset from the field of systems biology. Our results show that BN-GP can learn accurate Bayesian networks with significantly reduced computational cost compared to traditional methods. Furthermore, we provide a theoretical analysis of the convergence properties of BN-GP, which shows that it can learn consistent estimates of the network structure even in the presence of high-dimensional data. Our approach has the potential to enable the widespread adoption of Bayesian networks in high-dimensional systems, where traditional methods are often infeasible.\n\nKeywords: Bayesian networks, Gaussian process priors, uncertainty quantification, high-dimensional systems, systems biology.'</li></ul> | |
|
| Artificial Intelligence | <ul><li>'Title: Learning Hierarchical Representations for Robust Visual Perception in Autonomous Systems\nAbstract: We propose a novel deep learning approach for visual perception in autonomous systems, which leverages hierarchical representations to improve robustness and accuracy. Our method combines a convolutional neural network (CNN) with a recurrent neural network (RNN) to learn a hierarchical representation of visual data. We evaluate our approach on several benchmark datasets and demonstrate significant improvements in performance compared to state-of-the-art methods. Our results show that the proposed approach can learn robust and accurate representations of visual data, even in the presence of significant occlusions and variations in lighting conditions. We also provide a detailed analysis of the learned representations and demonstrate their applicability to various tasks in autonomous systems. This work makes significant contributions to the field of computer vision and robotics, and has the potential to enable more robust and accurate visual perception in autonomous systems.'</li><li>'Title: A Deep Learning Approach for Text Classification: A Comparative Study\n\nAbstract: Text classification is a fundamental task in natural language processing (NLP) that has numerous applications in various domains. In this paper, we propose a deep learning approach for text classification using convolutional neural networks (CNNs) and recurrent neural networks (RNNs). We compare the performance of our proposed approach with state-of-the-art methods on several benchmark datasets. Our results show that our approach outperforms the existing methods in terms of accuracy and F1-score. We also analyze the effect of different hyperparameters on the performance of our approach and provide insights into the importance of feature extraction in text classification. This study contributes to the development of efficient and accurate text classification models using deep learning techniques.\n\nKeywords: text classification, deep learning, convolutional neural networks, recurrent neural networks, natural language processing.'</li><li>'Title: Investigating the Impact of Attention Mechanisms on Deep Learning Models for Sentiment Analysis.\n\nAbstract: This paper explores the effects of incorporating attention mechanisms into deep learning models for sentiment analysis. We propose a novel architecture that combines the strengths of recurrent neural networks (RNNs) and attention mechanisms to improve the performance of sentiment analysis tasks. Our experimental results demonstrate that the proposed model outperforms state-of-the-art models in terms of accuracy and F1-score. Furthermore, we conduct an ablation study to investigate the impact of different attention mechanisms on the performance of the model. Our findings suggest that the proposed attention mechanism is more effective than other attention mechanisms in improving the performance of sentiment analysis tasks.\n\nKeywords: deep learning, attention mechanisms, sentiment analysis, natural language processing, neural networks.\n\nSource: Google Scholar.'</li></ul> | |
|
| Computational Engineering | <ul><li>'Title: An Efficient Seismic Inversion Method Using Deep Learning for Reservoir Characterization\nAbstract: Seismic inversion is a crucial step in reservoir characterization, and its accuracy directly affects the economic viability of hydrocarbon exploration and production. In this paper, we propose a novel seismic inversion method based on deep learning that can efficiently handle large-scale seismic data. The proposed method utilizes a convolutional neural network (CNN) to learn the mapping between seismic data and reservoir properties. We demonstrate the effectiveness of our method using a real-world dataset and show that it outperforms traditional methods in terms of accuracy and computational efficiency. Our results indicate that the proposed method can be a valuable tool for seismic inversion and reservoir characterization.\nKeywords: seismic inversion, deep learning, reservoir characterization, convolutional neural network, computational engineering.'</li><li>"A Novel Finite Element Method for Nonlinear Structural Analysis of Composite Materials\nAbstract: This paper presents a novel finite element method for nonlinear structural analysis of composite materials. The proposed method is based on a combination of the extended finite element method (XFEM) and the peridynamic theory (PDT). The XFEM is used to model the nonlinear behavior of the composite material, while the PDT is used to capture the long-range interactions between the material's particles. The proposed method is implemented in a computational framework and is validated using several numerical examples. The results show that the proposed method can accurately capture the nonlinear behavior of composite materials and can be used to predict the structural response of complex composite structures. The proposed method has the potential to be used in various engineering applications, including the design and analysis of composite structures for aerospace, automotive, and civil engineering."</li><li>'Title: A Novel Finite Element Method for Simulating Nonlinear Dynamics in Composite Materials\n\nAbstract: This paper presents a new finite element method for simulating nonlinear dynamics in composite materials. The proposed method combines the advantages of the partition of unity method and the extended finite element method to capture the complex behavior of composite materials under various loading conditions. The numerical results show that the proposed method can accurately predict the nonlinear dynamics of composite materials, including the effects of material nonlinearity and geometric nonlinearity. The proposed method is also compared with other existing methods, and the results show that it has better accuracy and efficiency.\n\nKeywords: finite element method, nonlinear dynamics, composite materials, partition of unity method, extended finite element method.\n\nArXiv ID: 2203.03045\n\nSubmission date: 2022-03-07\n\n'</li></ul> | |
|
| Computer Vision | <ul><li>'A Novel Approach to Object Detection using Convolutional Neural Networks\n\nAbstract: Object detection is a fundamental task in computer vision, and its applications are vast in various fields. In this paper, we propose a novel approach to object detection using convolutional neural networks (CNNs). Our method, called Object Detection using CNNs (ODCNN), is based on a combination of region proposal networks (RPNs) and CNNs. We train the ODCNN model on the PASCAL VOC 2007 dataset and evaluate its performance on the PASCAL VOC 2012 dataset. The results show that our approach outperforms the state-of-the-art methods in terms of accuracy and speed. We also provide a detailed analysis of the ODCNN model and its components. The code for the ODCNN model is available at https://github.com/odcnn/odcnn.\n\nKeywords: Object detection, Convolutional neural networks, Region proposal networks, PASCAL VOC 2007, PASCAL VOC 2012.'</li><li>'A Novel Object Recognition Framework for Autonomous Robots using Deep Learning and Computer Vision Techniques\n\nAbstract: This paper proposes a novel object recognition framework for autonomous robots that leverages the power of deep learning and computer vision techniques. The proposed framework consists of two stages: a detection stage and a recognition stage. In the detection stage, a convolutional neural network (CNN) is used to detect objects in the scene, while in the recognition stage, a recurrent neural network (RNN) is employed to recognize the detected objects. The proposed framework is evaluated on a dataset of images collected from a robotic platform, and the results show that it outperforms state-of-the-art methods in terms of accuracy and speed. The proposed framework has the potential to be used in various applications, including robotics, autonomous vehicles, and surveillance systems.\n\nKeywords: Object recognition, autonomous robots, deep learning, computer vision, convolutional neural networks, recurrent neural networks.\n\n'</li><li>'A novel approach to image classification using convolutional neural networks (CNNs) is proposed in this paper. The proposed method, dubbed "Deep Image Classifier", leverages the power of CNNs to learn hierarchical features from images. Experimental results on several benchmark datasets, including CIFAR-10 and ImageNet, demonstrate the efficacy of the proposed method in achieving state-of-the-art performance. The code for the proposed method is made available on GitHub, allowing for easy reproduction and extension of the results. The contributions of this paper can be summarized as follows: (1) a novel CNN architecture is proposed, which consists of multiple convolutional and pooling layers, followed by fully connected layers; (2) a novel training strategy is proposed, which involves data augmentation and batch normalization; (3) the proposed method is evaluated on several benchmark datasets, and the results are compared with state-of-the-art methods. The results of this paper demonstrate the potential of CNNs in image classification tasks, and provide a new benchmark for future research in this area.'</li></ul> | |
|
|
|
## Evaluation |
|
|
|
### Metrics |
|
| Label | F1 | |
|
|:--------|:-------| |
|
| **all** | 0.6184 | |
|
|
|
## Uses |
|
|
|
### Direct Use for Inference |
|
|
|
First install the SetFit library: |
|
|
|
```bash |
|
pip install setfit |
|
``` |
|
|
|
Then you can load this model and run inference. |
|
|
|
```python |
|
from setfit import SetFitModel |
|
|
|
# Download from the 🤗 Hub |
|
model = SetFitModel.from_pretrained("setfit_model_id") |
|
# Run inference |
|
preds = model("Let G be a finite group and let V be a finite-dimensional representation of G over an algebraically closed field k. We say that V is a representation of G in characteristic zero if the characteristic of k is zero. In this paper, we investigate the structure of the representation ring R(G) of a finite group G in characteristic zero. We show that R(G) is isomorphic to the group ring k[G] if and only if G is a cyclic group. Furthermore, we provide a characterization of the representation rings of finite abelian groups in terms of their irreducible representations. Our results have implications for the study of the representation theory of finite groups in characteristic zero.") |
|
``` |
|
|
|
<!-- |
|
### Downstream Use |
|
|
|
*List how someone could finetune this model on their own dataset.* |
|
--> |
|
|
|
<!-- |
|
### Out-of-Scope Use |
|
|
|
*List how the model may foreseeably be misused and address what users ought not to do with the model.* |
|
--> |
|
|
|
<!-- |
|
## Bias, Risks and Limitations |
|
|
|
*What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.* |
|
--> |
|
|
|
<!-- |
|
### Recommendations |
|
|
|
*What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.* |
|
--> |
|
|
|
## Training Details |
|
|
|
### Training Set Metrics |
|
| Training set | Min | Median | Max | |
|
|:-------------|:----|:---------|:-----| |
|
| Word count | 69 | 220.7380 | 1079 | |
|
|
|
| Label | Training Sample Count | |
|
|:--------------------------|:----------------------| |
|
| Commutative Algebra | 15 | |
|
| Computer Vision | 12 | |
|
| Artificial Intelligence | 16 | |
|
| Systems and Control | 19 | |
|
| Group Theory | 21 | |
|
| Computational Engineering | 16 | |
|
| Programming Languages | 13 | |
|
| Information Theory | 21 | |
|
| Data Structures | 21 | |
|
| Neural and Evolutionary | 21 | |
|
| Statistics Theory | 12 | |
|
|
|
### Training Hyperparameters |
|
- batch_size: (16, 16) |
|
- num_epochs: (5, 5) |
|
- max_steps: -1 |
|
- sampling_strategy: oversampling |
|
- body_learning_rate: (2e-05, 1e-05) |
|
- head_learning_rate: 0.01 |
|
- loss: CosineSimilarityLoss |
|
- distance_metric: cosine_distance |
|
- margin: 0.25 |
|
- end_to_end: False |
|
- use_amp: False |
|
- warmup_proportion: 0.1 |
|
- seed: 42 |
|
- eval_max_steps: -1 |
|
- load_best_model_at_end: True |
|
|
|
### Training Results |
|
| Epoch | Step | Training Loss | Validation Loss | |
|
|:-------:|:--------:|:-------------:|:---------------:| |
|
| 0.0005 | 1 | 0.158 | - | |
|
| 0.0253 | 50 | 0.1482 | - | |
|
| 0.0505 | 100 | 0.1408 | - | |
|
| 0.0758 | 150 | 0.1071 | - | |
|
| 0.1011 | 200 | 0.1294 | - | |
|
| 0.1263 | 250 | 0.0782 | - | |
|
| 0.1516 | 300 | 0.0628 | - | |
|
| 0.1769 | 350 | 0.0909 | - | |
|
| 0.2021 | 400 | 0.0161 | - | |
|
| 0.2274 | 450 | 0.0068 | - | |
|
| 0.2527 | 500 | 0.011 | - | |
|
| 0.2779 | 550 | 0.0027 | - | |
|
| 0.3032 | 600 | 0.0018 | - | |
|
| 0.3284 | 650 | 0.0011 | - | |
|
| 0.3537 | 700 | 0.0037 | - | |
|
| 0.3790 | 750 | 0.0015 | - | |
|
| 0.4042 | 800 | 0.0012 | - | |
|
| 0.4295 | 850 | 0.0006 | - | |
|
| 0.4548 | 900 | 0.0013 | - | |
|
| 0.4800 | 950 | 0.0004 | - | |
|
| 0.5053 | 1000 | 0.0003 | - | |
|
| 0.5306 | 1050 | 0.0001 | - | |
|
| 0.5558 | 1100 | 0.0007 | - | |
|
| 0.5811 | 1150 | 0.0001 | - | |
|
| 0.6064 | 1200 | 0.0004 | - | |
|
| 0.6316 | 1250 | 0.0001 | - | |
|
| 0.6569 | 1300 | 0.0001 | - | |
|
| 0.6822 | 1350 | 0.0006 | - | |
|
| 0.7074 | 1400 | 0.0002 | - | |
|
| 0.7327 | 1450 | 0.0002 | - | |
|
| 0.7580 | 1500 | 0.0001 | - | |
|
| 0.7832 | 1550 | 0.0001 | - | |
|
| 0.8085 | 1600 | 0.0001 | - | |
|
| 0.8338 | 1650 | 0.0001 | - | |
|
| 0.8590 | 1700 | 0.0002 | - | |
|
| 0.8843 | 1750 | 0.0001 | - | |
|
| 0.9096 | 1800 | 0.0001 | - | |
|
| 0.9348 | 1850 | 0.0001 | - | |
|
| 0.9601 | 1900 | 0.0001 | - | |
|
| 0.9853 | 1950 | 0.0001 | - | |
|
| **1.0** | **1979** | **-** | **0.0359** | |
|
| 1.0106 | 2000 | 0.0001 | - | |
|
| 1.0359 | 2050 | 0.0001 | - | |
|
| 1.0611 | 2100 | 0.0002 | - | |
|
| 1.0864 | 2150 | 0.0001 | - | |
|
| 1.1117 | 2200 | 0.0002 | - | |
|
| 1.1369 | 2250 | 0.0001 | - | |
|
| 1.1622 | 2300 | 0.0 | - | |
|
| 1.1875 | 2350 | 0.0003 | - | |
|
| 1.2127 | 2400 | 0.0001 | - | |
|
| 1.2380 | 2450 | 0.0001 | - | |
|
| 1.2633 | 2500 | 0.0001 | - | |
|
| 1.2885 | 2550 | 0.0 | - | |
|
| 1.3138 | 2600 | 0.0 | - | |
|
| 1.3391 | 2650 | 0.0001 | - | |
|
| 1.3643 | 2700 | 0.0046 | - | |
|
| 1.3896 | 2750 | 0.0044 | - | |
|
| 1.4149 | 2800 | 0.0005 | - | |
|
| 1.4401 | 2850 | 0.0002 | - | |
|
| 1.4654 | 2900 | 0.0001 | - | |
|
| 1.4907 | 2950 | 0.0 | - | |
|
| 1.5159 | 3000 | 0.0001 | - | |
|
| 1.5412 | 3050 | 0.0001 | - | |
|
| 1.5664 | 3100 | 0.0001 | - | |
|
| 1.5917 | 3150 | 0.0001 | - | |
|
| 1.6170 | 3200 | 0.0 | - | |
|
| 1.6422 | 3250 | 0.0 | - | |
|
| 1.6675 | 3300 | 0.0 | - | |
|
| 1.6928 | 3350 | 0.0 | - | |
|
| 1.7180 | 3400 | 0.0001 | - | |
|
| 1.7433 | 3450 | 0.0 | - | |
|
| 1.7686 | 3500 | 0.0 | - | |
|
| 1.7938 | 3550 | 0.0001 | - | |
|
| 1.8191 | 3600 | 0.0 | - | |
|
| 1.8444 | 3650 | 0.0 | - | |
|
| 1.8696 | 3700 | 0.0 | - | |
|
| 1.8949 | 3750 | 0.0 | - | |
|
| 1.9202 | 3800 | 0.0 | - | |
|
| 1.9454 | 3850 | 0.0 | - | |
|
| 1.9707 | 3900 | 0.0 | - | |
|
| 1.9960 | 3950 | 0.0 | - | |
|
| 2.0 | 3958 | - | 0.0579 | |
|
| 2.0212 | 4000 | 0.0 | - | |
|
| 2.0465 | 4050 | 0.0 | - | |
|
| 2.0718 | 4100 | 0.0001 | - | |
|
| 2.0970 | 4150 | 0.0001 | - | |
|
| 2.1223 | 4200 | 0.0 | - | |
|
| 2.1475 | 4250 | 0.0 | - | |
|
| 2.1728 | 4300 | 0.0 | - | |
|
| 2.1981 | 4350 | 0.0 | - | |
|
| 2.2233 | 4400 | 0.0 | - | |
|
| 2.2486 | 4450 | 0.0 | - | |
|
| 2.2739 | 4500 | 0.0 | - | |
|
| 2.2991 | 4550 | 0.0 | - | |
|
| 2.3244 | 4600 | 0.0001 | - | |
|
| 2.3497 | 4650 | 0.0 | - | |
|
| 2.3749 | 4700 | 0.0001 | - | |
|
| 2.4002 | 4750 | 0.0 | - | |
|
| 2.4255 | 4800 | 0.0 | - | |
|
| 2.4507 | 4850 | 0.0001 | - | |
|
| 2.4760 | 4900 | 0.0 | - | |
|
| 2.5013 | 4950 | 0.0 | - | |
|
| 2.5265 | 5000 | 0.0 | - | |
|
| 2.5518 | 5050 | 0.0 | - | |
|
| 2.5771 | 5100 | 0.0 | - | |
|
| 2.6023 | 5150 | 0.0 | - | |
|
| 2.6276 | 5200 | 0.0 | - | |
|
| 2.6529 | 5250 | 0.0 | - | |
|
| 2.6781 | 5300 | 0.0 | - | |
|
| 2.7034 | 5350 | 0.0001 | - | |
|
| 2.7287 | 5400 | 0.0 | - | |
|
| 2.7539 | 5450 | 0.0 | - | |
|
| 2.7792 | 5500 | 0.0001 | - | |
|
| 2.8044 | 5550 | 0.0 | - | |
|
| 2.8297 | 5600 | 0.0 | - | |
|
| 2.8550 | 5650 | 0.0 | - | |
|
| 2.8802 | 5700 | 0.0 | - | |
|
| 2.9055 | 5750 | 0.0 | - | |
|
| 2.9308 | 5800 | 0.0 | - | |
|
| 2.9560 | 5850 | 0.0 | - | |
|
| 2.9813 | 5900 | 0.0 | - | |
|
| 3.0 | 5937 | - | 0.0557 | |
|
| 3.0066 | 5950 | 0.0 | - | |
|
| 3.0318 | 6000 | 0.0 | - | |
|
| 3.0571 | 6050 | 0.0 | - | |
|
| 3.0824 | 6100 | 0.0 | - | |
|
| 3.1076 | 6150 | 0.0 | - | |
|
| 3.1329 | 6200 | 0.0 | - | |
|
| 3.1582 | 6250 | 0.0 | - | |
|
| 3.1834 | 6300 | 0.0 | - | |
|
| 3.2087 | 6350 | 0.0 | - | |
|
| 3.2340 | 6400 | 0.0 | - | |
|
| 3.2592 | 6450 | 0.0 | - | |
|
| 3.2845 | 6500 | 0.0 | - | |
|
| 3.3098 | 6550 | 0.0 | - | |
|
| 3.3350 | 6600 | 0.0 | - | |
|
| 3.3603 | 6650 | 0.0 | - | |
|
| 3.3855 | 6700 | 0.0 | - | |
|
| 3.4108 | 6750 | 0.0 | - | |
|
| 3.4361 | 6800 | 0.0 | - | |
|
| 3.4613 | 6850 | 0.0 | - | |
|
| 3.4866 | 6900 | 0.0 | - | |
|
| 3.5119 | 6950 | 0.0 | - | |
|
| 3.5371 | 7000 | 0.0 | - | |
|
| 3.5624 | 7050 | 0.0 | - | |
|
| 3.5877 | 7100 | 0.0 | - | |
|
| 3.6129 | 7150 | 0.0 | - | |
|
| 3.6382 | 7200 | 0.0 | - | |
|
| 3.6635 | 7250 | 0.0 | - | |
|
| 3.6887 | 7300 | 0.0 | - | |
|
| 3.7140 | 7350 | 0.0 | - | |
|
| 3.7393 | 7400 | 0.0 | - | |
|
| 3.7645 | 7450 | 0.0 | - | |
|
| 3.7898 | 7500 | 0.0 | - | |
|
| 3.8151 | 7550 | 0.0 | - | |
|
| 3.8403 | 7600 | 0.0 | - | |
|
| 3.8656 | 7650 | 0.0 | - | |
|
| 3.8909 | 7700 | 0.0 | - | |
|
| 3.9161 | 7750 | 0.0 | - | |
|
| 3.9414 | 7800 | 0.0 | - | |
|
| 3.9666 | 7850 | 0.0 | - | |
|
| 3.9919 | 7900 | 0.0 | - | |
|
| 4.0 | 7916 | - | 0.0543 | |
|
| 4.0172 | 7950 | 0.0 | - | |
|
| 4.0424 | 8000 | 0.0 | - | |
|
| 4.0677 | 8050 | 0.0 | - | |
|
| 4.0930 | 8100 | 0.0 | - | |
|
| 4.1182 | 8150 | 0.0 | - | |
|
| 4.1435 | 8200 | 0.0 | - | |
|
| 4.1688 | 8250 | 0.0 | - | |
|
| 4.1940 | 8300 | 0.0 | - | |
|
| 4.2193 | 8350 | 0.0 | - | |
|
| 4.2446 | 8400 | 0.0 | - | |
|
| 4.2698 | 8450 | 0.0 | - | |
|
| 4.2951 | 8500 | 0.0 | - | |
|
| 4.3204 | 8550 | 0.0 | - | |
|
| 4.3456 | 8600 | 0.0 | - | |
|
| 4.3709 | 8650 | 0.0 | - | |
|
| 4.3962 | 8700 | 0.0 | - | |
|
| 4.4214 | 8750 | 0.0 | - | |
|
| 4.4467 | 8800 | 0.0 | - | |
|
| 4.4720 | 8850 | 0.0 | - | |
|
| 4.4972 | 8900 | 0.0 | - | |
|
| 4.5225 | 8950 | 0.0 | - | |
|
| 4.5478 | 9000 | 0.0 | - | |
|
| 4.5730 | 9050 | 0.0 | - | |
|
| 4.5983 | 9100 | 0.0 | - | |
|
| 4.6235 | 9150 | 0.0 | - | |
|
| 4.6488 | 9200 | 0.0 | - | |
|
| 4.6741 | 9250 | 0.0 | - | |
|
| 4.6993 | 9300 | 0.0 | - | |
|
| 4.7246 | 9350 | 0.0 | - | |
|
| 4.7499 | 9400 | 0.0 | - | |
|
| 4.7751 | 9450 | 0.0 | - | |
|
| 4.8004 | 9500 | 0.0 | - | |
|
| 4.8257 | 9550 | 0.0 | - | |
|
| 4.8509 | 9600 | 0.0 | - | |
|
| 4.8762 | 9650 | 0.0 | - | |
|
| 4.9015 | 9700 | 0.0 | - | |
|
| 4.9267 | 9750 | 0.0 | - | |
|
| 4.9520 | 9800 | 0.0 | - | |
|
| 4.9773 | 9850 | 0.0 | - | |
|
| 5.0 | 9895 | - | 0.0537 | |
|
|
|
* The bold row denotes the saved checkpoint. |
|
### Framework Versions |
|
- Python: 3.9.19 |
|
- SetFit: 1.1.0.dev0 |
|
- Sentence Transformers: 3.0.1 |
|
- Transformers: 4.39.0 |
|
- PyTorch: 2.4.0 |
|
- Datasets: 2.20.0 |
|
- Tokenizers: 0.15.2 |
|
|
|
## Citation |
|
|
|
### BibTeX |
|
```bibtex |
|
@article{https://doi.org/10.48550/arxiv.2209.11055, |
|
doi = {10.48550/ARXIV.2209.11055}, |
|
url = {https://arxiv.org/abs/2209.11055}, |
|
author = {Tunstall, Lewis and Reimers, Nils and Jo, Unso Eun Seo and Bates, Luke and Korat, Daniel and Wasserblat, Moshe and Pereg, Oren}, |
|
keywords = {Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences}, |
|
title = {Efficient Few-Shot Learning Without Prompts}, |
|
publisher = {arXiv}, |
|
year = {2022}, |
|
copyright = {Creative Commons Attribution 4.0 International} |
|
} |
|
``` |
|
|
|
<!-- |
|
## Glossary |
|
|
|
*Clearly define terms in order to be accessible across audiences.* |
|
--> |
|
|
|
<!-- |
|
## Model Card Authors |
|
|
|
*Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.* |
|
--> |
|
|
|
<!-- |
|
## Model Card Contact |
|
|
|
*Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.* |
|
--> |