EmpiricalLearnerLoss.java

  1. package org.drip.learning.bound;

  2. /*
  3.  * -*- mode: java; tab-width: 4; indent-tabs-mode: nil; c-basic-offset: 4 -*-
  4.  */

  5. /*!
  6.  * Copyright (C) 2020 Lakshmi Krishnamurthy
  7.  * Copyright (C) 2019 Lakshmi Krishnamurthy
  8.  * Copyright (C) 2018 Lakshmi Krishnamurthy
  9.  * Copyright (C) 2017 Lakshmi Krishnamurthy
  10.  * Copyright (C) 2016 Lakshmi Krishnamurthy
  11.  * Copyright (C) 2015 Lakshmi Krishnamurthy
  12.  *
  13.  *  This file is part of DROP, an open-source library targeting analytics/risk, transaction cost analytics,
  14.  *      asset liability management analytics, capital, exposure, and margin analytics, valuation adjustment
  15.  *      analytics, and portfolio construction analytics within and across fixed income, credit, commodity,
  16.  *      equity, FX, and structured products. It also includes auxiliary libraries for algorithm support,
  17.  *      numerical analysis, numerical optimization, spline builder, model validation, statistical learning,
  18.  *      and computational support.
  19.  *  
  20.  *      https://lakshmidrip.github.io/DROP/
  21.  *  
  22.  *  DROP is composed of three modules:
  23.  *  
  24.  *  - DROP Product Core - https://lakshmidrip.github.io/DROP-Product-Core/
  25.  *  - DROP Portfolio Core - https://lakshmidrip.github.io/DROP-Portfolio-Core/
  26.  *  - DROP Computational Core - https://lakshmidrip.github.io/DROP-Computational-Core/
  27.  *
  28.  *  DROP Product Core implements libraries for the following:
  29.  *  - Fixed Income Analytics
  30.  *  - Loan Analytics
  31.  *  - Transaction Cost Analytics
  32.  *
  33.  *  DROP Portfolio Core implements libraries for the following:
  34.  *  - Asset Allocation Analytics
  35.  *  - Asset Liability Management Analytics
  36.  *  - Capital Estimation Analytics
  37.  *  - Exposure Analytics
  38.  *  - Margin Analytics
  39.  *  - XVA Analytics
  40.  *
  41.  *  DROP Computational Core implements libraries for the following:
  42.  *  - Algorithm Support
  43.  *  - Computation Support
  44.  *  - Function Analysis
  45.  *  - Model Validation
  46.  *  - Numerical Analysis
  47.  *  - Numerical Optimizer
  48.  *  - Spline Builder
  49.  *  - Statistical Learning
  50.  *
  51.  *  Documentation for DROP is Spread Over:
  52.  *
  53.  *  - Main                     => https://lakshmidrip.github.io/DROP/
  54.  *  - Wiki                     => https://github.com/lakshmiDRIP/DROP/wiki
  55.  *  - GitHub                   => https://github.com/lakshmiDRIP/DROP
  56.  *  - Repo Layout Taxonomy     => https://github.com/lakshmiDRIP/DROP/blob/master/Taxonomy.md
  57.  *  - Javadoc                  => https://lakshmidrip.github.io/DROP/Javadoc/index.html
  58.  *  - Technical Specifications => https://github.com/lakshmiDRIP/DROP/tree/master/Docs/Internal
  59.  *  - Release Versions         => https://lakshmidrip.github.io/DROP/version.html
  60.  *  - Community Credits        => https://lakshmidrip.github.io/DROP/credits.html
  61.  *  - Issues Catalog           => https://github.com/lakshmiDRIP/DROP/issues
  62.  *  - JUnit                    => https://lakshmidrip.github.io/DROP/junit/index.html
  63.  *  - Jacoco                   => https://lakshmidrip.github.io/DROP/jacoco/index.html
  64.  *
  65.  *  Licensed under the Apache License, Version 2.0 (the "License");
  66.  *      you may not use this file except in compliance with the License.
  67.  *  
  68.  *  You may obtain a copy of the License at
  69.  *      http://www.apache.org/licenses/LICENSE-2.0
  70.  *  
  71.  *  Unless required by applicable law or agreed to in writing, software
  72.  *      distributed under the License is distributed on an "AS IS" BASIS,
  73.  *      WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  74.  *  
  75.  *  See the License for the specific language governing permissions and
  76.  *      limitations under the License.
  77.  */

  78. /**
  79.  * <i>EmpiricalLearnerLoss</i> Function computes the Empirical Loss of a Learning Operation resulting from
  80.  * the Use of a Learning Function in Conjunction with the corresponding Empirical Realization. The References
  81.  * are:
  82.  * <br><br>
  83.  * <ul>
  84.  *  <li>
  85.  *      Boucheron, S., G. Lugosi, and P. Massart (2003): Concentration Inequalities Using the Entropy Method
  86.  *          <i>Annals of Probability</i> <b>31</b> 1583-1614
  87.  *  </li>
  88.  *  <li>
  89.  *      Lugosi, G. (2002): Pattern Classification and Learning Theory, in: <i>L. Györ, editor, Principles
  90.  *          of Non-parametric Learning</i> <b>Springer</b> Wien 5-62
  91.  *  </li>
  92.  * </ul>
  93.  *
  94.  *  <br><br>
  95.  *  <ul>
  96.  *      <li><b>Module </b> = <a href = "https://github.com/lakshmiDRIP/DROP/tree/master/ComputationalCore.md">Computational Core Module</a></li>
  97.  *      <li><b>Library</b> = <a href = "https://github.com/lakshmiDRIP/DROP/tree/master/StatisticalLearningLibrary.md">Statistical Learning</a></li>
  98.  *      <li><b>Project</b> = <a href = "https://github.com/lakshmiDRIP/DROP/tree/master/src/main/java/org/drip/learning">Agnostic Learning Bounds under Empirical Loss Minimization Schemes</a></li>
  99.  *      <li><b>Package</b> = <a href = "https://github.com/lakshmiDRIP/DROP/tree/master/src/main/java/org/drip/learning/bound">Covering Numbers, Concentration, Lipschitz Bounds</a></li>
  100.  *  </ul>
  101.  *
  102.  * @author Lakshmi Krishnamurthy
  103.  */

  104. public class EmpiricalLearnerLoss extends org.drip.function.definition.R1ToR1 {
  105.     private double _dblRealization = java.lang.Double.NaN;
  106.     private org.drip.function.definition.R1ToR1 _learner = null;

  107.     /**
  108.      * EmpiricalLearnerLoss Constructor
  109.      *
  110.      * @param learner The Learning Function
  111.      * @param dblRealization The Empirical Outcome
  112.      *
  113.      * @throws java.lang.Exception Thrown if the Inputs are Invalid
  114.      */

  115.     public EmpiricalLearnerLoss (
  116.         final org.drip.function.definition.R1ToR1 learner,
  117.         final double dblRealization)
  118.         throws java.lang.Exception
  119.     {
  120.         super (null);

  121.         if (null == (_learner = learner) || !org.drip.numerical.common.NumberUtil.IsValid (_dblRealization =
  122.             dblRealization))
  123.             throw new java.lang.Exception ("EmpiricalLearnerLoss ctr: Invalid Inputs");
  124.     }

  125.     /**
  126.      * Retrieve the Empirical Realization
  127.      *
  128.      * @return The Empirical Realization
  129.      */

  130.     public double empiricalRealization()
  131.     {
  132.         return _dblRealization;
  133.     }

  134.     /**
  135.      * Retrieve the Learning Function
  136.      *
  137.      * @return The Learning Function
  138.      */

  139.     public org.drip.function.definition.R1ToR1 learner()
  140.     {
  141.         return _learner;
  142.     }

  143.     /**
  144.      * Compute the Loss for the specified Variate
  145.      *
  146.      * @param dblVariate The Variate
  147.      *
  148.      * @return Loss for the specified Variate
  149.      *
  150.      * @throws java.lang.Exception Thrown if the Loss cannot be computed
  151.      */

  152.     public double loss (
  153.         final double dblVariate)
  154.         throws java.lang.Exception
  155.     {
  156.         return _dblRealization - _learner.evaluate (dblVariate);
  157.     }

  158.     @Override public double evaluate (
  159.         final double dblVariate)
  160.         throws java.lang.Exception
  161.     {
  162.         return loss (dblVariate);
  163.     }
  164. }