[PDF] ctree : Conditional Inference Trees | Semantic Scholar (2024)

Skip to search formSkip to main contentSkip to account menu

Semantic ScholarSemantic Scholar's Logo
  • Corpus ID: 3730942
@inproceedings{Hothorn2015ctreeC, title={ctree : Conditional Inference Trees}, author={Torsten Hothorn and Kurt Hornik and Wirtschaftsuniversit{\"a}t Wien and Achim Zeileis}, year={2015}, url={https://api.semanticscholar.org/CorpusID:3730942}}
  • T. Hothorn, K. Hornik, A. Zeileis
  • Published 2015
  • Computer Science, Mathematics

This vignette describes the new reimplementation of conditional inference trees (CTree) in the R package partykit . CTree is a non-parametric class of regression trees embedding tree-structured

81 Citations

Highly Influential Citations

5

Background Citations

14

Methods Citations

34

Figures from this paper

  • figure 1
  • figure 2
  • figure 3
  • figure 4
  • figure 5
  • figure 6
  • figure 7
  • figure 8
  • figure 9

Topics

Conditional Inference Tree (opens in a new tab)Partykit (opens in a new tab)Inference Procedure (opens in a new tab)Censored (opens in a new tab)Regression Trees (opens in a new tab)Tree-structured Regression Models (opens in a new tab)Covariates (opens in a new tab)R Package (opens in a new tab)Measurement Scales (opens in a new tab)

81 Citations

MCCE: Monte Carlo sampling of realistic counterfactual explanations
    Annabelle RedelmeierMartin JullumK. AasAnders Løland

    Computer Science

    ArXiv

  • 2021

MCCE is introduced, a novel counterfactual explanation method that generates on-manifold, actionable and valid counterfactuals by modeling the joint distribution of the mutable features given the immutable features and the decision by modeling the joint distribution of the mutable features given the immutable features and the decision.

A survival tree based on stabilized score tests for high-dimensional covariates
    T. EmuraWei-Chern HsuW. Chou

    Mathematics, Medicine

    Journal of applied statistics

  • 2023

This work proposes a novel matrix-based algorithm in order to tests a number of nodes simultaneously via stabilized score tests and proposes a recursive partitioning algorithm to construct a survival tree and develops the original R package uni.survival.tree (https://cran.r-project.org/package=uni. survival.tree) for implementation.

  • 4
  • PDF
Comparing Variable Importance in Prediction of Silence Behaviours between Random Forest and Conditional Inference Forest Models.
    Stephen BarrettG. GrayColm McGuinnessM. Knoll

    Computer Science, Sociology

  • 2020

The models analysed explored the role of cultural factors at individual and societal level when predicting Organisational Silence behaviours and why CIT should be used when dealing with data with different levels of aggregation.

  • 1
  • PDF
Finding Process Variants in Event Logs
    Alfredo Bolt

    Computer Science

  • 2017

This paper introduces an unsupervised and generic technique to detect significant variants in event logs by applying existing, well-proven data mining techniques for recursive partitioning driven by conditional inference over event attributes.

  • 16
  • PDF
Survival trees based on heterogeneity in time‐to‐event and censoring distributions using parameter instability test
    M. G. KunduSamiran Ghosh

    Mathematics, Medicine

    Stat. Anal. Data Min.

  • 2021

The proposed SurvCART algorithm utilizes the “conditional inference” framework that selects splitting variable via parameter instability test and subsequently finds the optimal split based on some maximally chosen statistic.

Interpretable Machine Learning - A Brief History, State-of-the-Art and Challenges
    Christoph MolnarGiuseppe CasalicchioB. Bischl

    Computer Science

    PKDD/ECML Workshops

  • 2020

The field is urged to recall its roots of interpretable, data-driven modeling in statistics and (rule-based) ML, but also to consider other areas such as sensitivity analysis, causal inference, and the social sciences.

Locating disparities in machine learning
    Moritz von ZahnO. HinzS. Feuerriegel

    Computer Science

    2023 IEEE International Conference on Big Data…

  • 2023

This work proposes a data-driven framework called Automatic Location of Disparities (ALD), which aims at locating disparities in machine learning and produces interpretable audit reports as output and demonstrates the effectiveness of ALD based on both synthetic and real-world datasets.

DecisionTree for Classification and Regression: A State-of-the Art Review
    M. JenaS. Dehuri

    Computer Science, Mathematics

    Informatica

  • 2020

This paper reviews extensively many popularly used state-of-the-art decision tree-based techniques for classification and regression and presents a survey of morethan forty years of research that has been emphasized on the application of decision trees in bothclassification and regression.

  • 17
  • PDF
Personalization of Medical Treatment Decisions: Simplifying Complex Models while Maintaining Patient Health Outcomes
    Christopher WeyantM. Brandeau

    Medicine, Computer Science

    Medical decision making : an international…

  • 2021

The meta-modeling method is disease- and model- agnostic and can be used to simplify complex models for personalization, allowing for variable selection in addition to improved model interpretability and computational performance.

  • 7
  • PDF
DNN Explanation for Safety Analysis: an Empirical Evaluation of Clustering-based Approaches
    M. AttaouiHazem M. FahmyF. PastoreLionel C. Briand

    Computer Science, Engineering

    ArXiv

  • 2023

An empirical evaluation of 99 different pipelines for root cause analysis of DNN failures shows that the best pipeline combines transfer learning, DBSCAN, and UMAP and generates distinct clusters for each root cause of failure, thus enabling engineers to detect all the unsafe scenarios.

  • 1
  • PDF

...

...

25 References

Unbiased Recursive Partitioning: A Conditional Inference Framework
    T. HothornK. HornikA. Zeileis

    Mathematics, Computer Science

  • 2006

A unified framework for recursive partitioning is proposed which embeds tree-structured regression models into a well defined theory of conditional inference procedures and it is shown that the predicted accuracy of trees with early stopping is equivalent to the prediction accuracy of pruned trees with unbiased variable selection.

  • 3,316
  • PDF
Regression Trees for Censored Data
    M. Segal

    Mathematics

  • 1988

The regression-tree methodology is extended to right-censored response variables by replacing the conventional splitting rules with rules based on the Tarone-Ware or Harrington-Fleming classes of

  • 471
  • PDF
Categorical Data Analysis
    Jeremy FreeseJason Beckfield

    Mathematics, Computer Science

  • 2001

This workshop introduces students to current methods for analyzing categorical data, with its principal focus being regression models for categorical outcomes. We will consider models for binary,

  • 13,015
  • PDF
Relative risk trees for censored survival data.
    M. LeBlancJ. Crowley

    Mathematics

    Biometrics

  • 1992

A method is developed for obtaining tree-structured relative risk estimates for censored survival data using a recursive partitioning algorithm that adopts most aspects of the widely used Classification and Regression Tree (CART) algorithm.

  • 421
Tree-based multivariate regression and density estimation with right-censored data
    A. MolinaroS. DudoitM. J. Laan

    Mathematics

  • 2004
  • 89
  • PDF
Applied Logistic Regression
    Joseph D. Conklin

    Mathematics

    Technometrics

  • 2002

As a consultant, I am always on the lookout for new books that help me do my job better. Iwould recommend practitioners of regression, that is, probably most of us, to read and use this book. Anthony

  • 2,473
Bias in information-based measures in decision tree induction
    A. WhiteWei Zhong Liu

    Computer Science, Mathematics

    Machine Learning

  • 2004

A fresh look is taken at the problem of bias in information-based attribute selection measures, used in the induction of decision trees and it is concluded that approaches which utilise the chi-square distribution are preferable because they compensate automatically for differences between attributes in the number of levels they take.

  • 158
  • PDF
Classification and regression trees
    N. Speybroeck

    Medicine, Computer Science

    International Journal of Public Health

  • 2011

As with stepwise linear regression procedures, adding variables will continuously increase the fit of the model to the data, but at the cost of increasing the true fit to an independent data set.

  • 12,942
  • Highly Influential
An unbiased method for constructing multilabel classification trees
    Hyun Gon NohM. SongSung Hyun Park

    Computer Science, Mathematics

    Comput. Stat. Data Anal.

  • 2004
  • 24
On the Asymptotic Theory of Permutation Statistics
    H. StrasserChristian H. Weber

    Mathematics

  • 1999

In this paper limit theorems for the conditional distributions of linear test statistics are proved. The assertions are conditioned by the sigma-field of permutation symmetric sets. Limit theorems

  • 295
  • Highly Influential
  • PDF

...

...

Related Papers

Showing 1 through 3 of 0 Related Papers

    [PDF] ctree : Conditional Inference Trees | Semantic Scholar (2024)

    FAQs

    What are conditional inference trees? ›

    Conditional inference trees estimate a regression relationship by binary recursive partitioning in a conditional inference framework. Roughly, the algorithm works as follows: 1) Test the global null hypothesis of independence between any of the input variables and the response (which may be multivariate as well).

    What is CTree in R? ›

    CTree is a non-parametric class of regression trees embedding tree-structured regression models into a well defined theory of conditional inference pro- cedures.

    What is a conditional random forest? ›

    A CRF is an ensemble of multiple CITs. The algorithm uses resampling with or without replacement to create a random sample for each tree. Importantly, only a sample of candidate predictors is randomly drawn for each individual CITs.

    How to build a classification tree in R? ›

    How to build classification trees in R?
    1. Recipe Objective. ...
    2. STEP 1: Importing Necessary Libraries. ...
    3. STEP 2: Loading the Train and Test Dataset. ...
    4. STEP 3: Data Preprocessing (Scaling) ...
    5. STEP 4: Creation of Decision Tree Classifier model using training set. ...
    6. STEP 5: Predict using Test Dataset. ...
    7. STEP 6: Creation of confusion matrix.
    Dec 26, 2022

    What is the difference between CIT and cart? ›

    Conditional Inference Trees (CITs) are much better at determining the true effect of a predictor, i.e. the effect of a predictor if all other effects are simultaneously considered. In contrast to CARTs, CITs use p-values to determine splits in the data.

    What is a limitation of decision trees? ›

    One of the limitations of decision trees is that they are largely unstable compared to other decision predictors. A small change in the data can result in a major change in the structure of the decision tree, which can convey a different result from what users will get in a normal event.

    What is the difference between MRF and CRF? ›

    MRF and CRF share the same graphical models, but MRF are generative models which model the joint probability distribution, while CRF are discriminative models which model the conditional probability distribution.

    What is a conditional probability tree? ›

    Often we use tree diagrams to model conditional probability. This is where there is more than one outcome and they are not independent – in other words the first outcome affects the probability of the second.

    What is the difference between a random forest and a tree? ›

    Note that the random forest is a predictive modeling tool, not a descriptive one. The random forest has complex data visualization and accurate predictions, but the decision tree has simple visualization and less accurate predictions.

    Can random forest do classification? ›

    Random forest is a flexible, easy-to-use machine learning algorithm that produces, even without hyper-parameter tuning, a great result most of the time. It is also one of the most-used algorithms, due to its simplicity and diversity (it can be used for both classification and regression tasks).

    How to predict using a decision tree? ›

    A. A decision tree algorithm is a machine learning algorithm that uses a decision tree to make predictions. It follows a tree-like model of decisions and their possible consequences. The algorithm works by recursively splitting the data into subsets based on the most significant feature at each node of the tree.

    What is the difference between R * tree and R tree? ›

    In data processing R*-trees are a variant of R-trees used for indexing spatial information. R*-trees have slightly higher construction cost than standard R-trees, as the data may need to be reinserted; but the resulting tree will usually have a better query performance.

    What are decision trees for inference? ›

    Inferring a decision tree from a given dataset is a classic problem in machine learning. This problem consists of building, from a labelled dataset, a tree where each node corresponds to a class and a path between the tree root and a leaf corresponds to a conjunction of features to be satisfied in this class.

    What are conditions in decision trees? ›

    Conditions with two possible outcomes (for example, true or false) are called binary conditions. Decision trees containing only binary conditions are called binary decision trees. Non-binary conditions have more than two possible outcomes.

    What is an example of a decision tree? ›

    A decision tree is a tree-like structure that represents a series of decisions and their possible consequences. It is used in machine learning for classification and regression tasks. An example of a decision tree is a flowchart that helps a person decide what to wear based on the weather conditions.

    What are decision trees in causal inference? ›

    Decision trees for causal inference are generally used to separate data into buckets in order to estimate the average treatment effects within each node.

    Top Articles
    Latest Posts
    Article information

    Author: Nathanial Hackett

    Last Updated:

    Views: 5663

    Rating: 4.1 / 5 (52 voted)

    Reviews: 91% of readers found this page helpful

    Author information

    Name: Nathanial Hackett

    Birthday: 1997-10-09

    Address: Apt. 935 264 Abshire Canyon, South Nerissachester, NM 01800

    Phone: +9752624861224

    Job: Forward Technology Assistant

    Hobby: Listening to music, Shopping, Vacation, Baton twirling, Flower arranging, Blacksmithing, Do it yourself

    Introduction: My name is Nathanial Hackett, I am a lovely, curious, smiling, lively, thoughtful, courageous, lively person who loves writing and wants to share my knowledge and understanding with you.