flink-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "ASF GitHub Bot (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (FLINK-1979) Implement Loss Functions
Date Tue, 17 May 2016 17:01:12 GMT

    [ https://issues.apache.org/jira/browse/FLINK-1979?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15287013#comment-15287013
] 

ASF GitHub Bot commented on FLINK-1979:
---------------------------------------

Github user skavulya commented on a diff in the pull request:

    https://github.com/apache/flink/pull/1985#discussion_r63561773
  
    --- Diff: flink-libraries/flink-ml/src/main/scala/org/apache/flink/ml/optimization/GradientDescent.scala
---
    @@ -272,7 +272,7 @@ abstract class GradientDescent extends IterativeSolver {
       * The regularization function is `1/2 ||w||_2^2` with `w` being the weight vector.
       */
     class GradientDescentL2 extends GradientDescent {
    -
    +  //TODO(skavulya): Pass regularization penalty as a parameter
    --- End diff --
    
    I can add it to this PR since it is only affecting SGD.


> Implement Loss Functions
> ------------------------
>
>                 Key: FLINK-1979
>                 URL: https://issues.apache.org/jira/browse/FLINK-1979
>             Project: Flink
>          Issue Type: Improvement
>          Components: Machine Learning Library
>            Reporter: Johannes Günther
>            Assignee: Johannes Günther
>            Priority: Minor
>              Labels: ML
>
> For convex optimization problems, optimizer methods like SGD rely on a pluggable implementation
of a loss function and its first derivative.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Mime
View raw message