Model Selection

497 papers with code • 0 benchmarks • 1 datasets

Given a set of candidate models, the goal of Model Selection is to select the model that best approximates the observed data and captures its underlying regularities. Model Selection criteria are defined such that they strike a balance between the goodness of fit, and the generalizability or complexity of the models.

Source: Kernel-based Information Criterion

Libraries

Use these libraries to find Model Selection models and implementations

Evaluating Large Language Models as Generative User Simulators for Conversational Recommendation

granelle/naacl24-user-sim 13 Mar 2024

Synthetic users are cost-effective proxies for real users in the evaluation of conversational recommender systems.

4
13 Mar 2024

Detection of Unobserved Common Causes based on NML Code in Discrete, Mixed, and Continuous Variables

matsushima-lab/cloud 11 Mar 2024

Causal discovery in the presence of unobserved common causes from observational data only is a crucial but challenging problem.

0
11 Mar 2024

On the Model-Agnostic Multi-Source-Free Unsupervised Domain Adaptation

spiresearch/mmda 3 Mar 2024

Specifically, we first conduct source model selection based on the proposed selection principles.

0
03 Mar 2024

Defining Expertise: Applications to Treatment Effect Estimation

qiyaowei/expertise 1 Mar 2024

Actions of an expert thus naturally encode part of their domain knowledge, and can help make inferences within the same domain: Knowing doctors try to prescribe the best treatment for their patients, we can tell treatments prescribed more frequently are likely to be more effective.

0
01 Mar 2024

Benchmark Self-Evolving: A Multi-Agent Framework for Dynamic LLM Evaluation

nanshineloong/self-evolving-benchmark 18 Feb 2024

Towards a more scalable, robust and fine-grained evaluation, we implement six reframing operations to construct evolving instances testing LLMs against diverse queries, data noise and probing their problem-solving sub-abilities.

6
18 Feb 2024

LoTa-Bench: Benchmarking Language-oriented Task Planners for Embodied Agents

lbaa2022/llmtaskplanning 13 Feb 2024

Large language models (LLMs) have recently received considerable attention as alternative solutions for task planning.

32
13 Feb 2024

Model Assessment and Selection under Temporal Distribution Shift

eliselyhan/arw 13 Feb 2024

We investigate model assessment and selection in a changing environment, by synthesizing datasets from both the current time period and historical epochs.

1
13 Feb 2024

MT-HCCAR: Multi-Task Deep Learning with Hierarchical Classification and Attention-based Regression for Cloud Property Retrieval

ai-4-atmosphere-remote-sensing/mt-hccar 29 Jan 2024

In response, this paper introduces MT-HCCAR, an end-to-end deep learning model employing multi-task learning to simultaneously tackle cloud masking, cloud phase retrieval (classification tasks), and COT prediction (a regression task).

3
29 Jan 2024

INTERS: Unlocking the Power of Large Language Models in Search with Instruction Tuning

daod/inters 12 Jan 2024

Despite this, their application to information retrieval (IR) tasks is still challenging due to the infrequent occurrence of many IR-specific concepts in natural language.

184
12 Jan 2024

Valid causal inference with unobserved confounding in high-dimensional settings

stat4reg/hdim.ui 12 Jan 2024

We propose uncertainty intervals which allow for unobserved confounding, and show that the resulting inference is valid when the amount of unobserved confounding is small relative to the sample size; the latter is formalized in terms of convergence rates.

0
12 Jan 2024