Summary and Contributions: After reading the authors' response, I remain positive about the paper. === This paper studies online DR-submodular maximization problemn with convex constraints. The authors proposed an algorithm that achieves sublinear regret and sublinear constraint violation under both stochastic and adversarial models.
Strengths: The authors proposed an unified approach that enjoys theoretical guarantees in both adversarial and stochastic settings.
Weaknesses: This paper may lack discussion on lower bounds for the problem in convex and DR-submodular setting. The authors are recommended to discuss the optimality/suboptimality of the previous works and the proposed algorithm. Fig 1 has several issues. 1) The fonts look very tiny and readers have to zoom in. 2) It would be better to give names to Alg 1 and 2 so that future works can refer to them conveniently. 3) I see a red vertical line (Alg 2) in fig1b. It looks a little bizarre. 4) Is there any reason that can account for the decline of Alg 1 in fig1b starting from t=0.3? For the experiment section, the authors may want to consider comparing the proposed algorithm to Meta-Frank-Wolfe, although it is not designed for a problem with a violation constraint.
Correctness: I didn't check the proofs in the appendix and the main body looks correct to me.
Relation to Prior Work: Yes
Summary and Contributions: In this paper, the author(s) studied the online DR-submodular maximization problem, under adversarial and stochastic settings, with convex constraints. Two algorithms are proposed, and both sublinear regret bound and sublinear total constraint violation bound are established.
Strengths: This work is believed to be the first work which achieves sublinear bound for cumulative constraint under a general convex long-term constraints. The regret bound is established in terms of both expectation and high probability. In order to obtain more efficiency, a faster algorithms is proposed, too. There are also plenty of empirical experiments (e.g., online joke recommendation, online task assignment in crowdsourcing markets, online welfare maximization with production cost), which validate the performance of the proposed algorithms.
Weaknesses: It seems that the proposed algorithms are straightforward combination of existing online submodular maximisation algorithms ([18, 20]) and the algorithm for online convex problem with adversarial cumulative constraints (). So the novelty of this work should be highlighted. For example, what is the difficulty when we try to combine these two methods, and how the author(s) circumvented the obstacles. Please also emphasize the new ideas/techniques utilized in the work.
Correctness: The claims and results look correct to me, although I didn’t check the proof carefully.
Clarity: This paper is well-written and easy to follow. One suggestion: in line 172, the definition of x_t^* should be provided.
Relation to Prior Work: The motivation and contributions of this work (compared with the existing works) are well explained and highlighted.
Additional Feedback: My major concern is the novelty of this work. The ideas and techniques are not sufficiently original to be accepted by NeurIPS. ------ Update: The author(s) addressed my concerns about the novelty in the response. So I raised the score to "marginally above the acceptance threshold".
Summary and Contributions: The paper considers the problem of maximizing a general monotone DR-submodular function subject to a general convex constraint (general up to some natural assumptions) in the online regret-minimization setting. The paper presents two algorithms for this problem, and proves bounds on their regret (with respect to the 1-1/e offline approximation) as well as the extent to which they violate the constraint on average. In that respect, the paper considers three kinds of regrets: - The traditional adversarial static regret in which the input is selected by an adversary and the algorithm competes with the best single solution in hindsight. - An adversarial dynamic regret in which the input is selected by an adversary and the algorithm competes with the best offline algorithm. - A stochastic static regret in which the constraint in every round is chosen i.i.d. from some unknown distribution, and the algorithm competes with the best single solution that obeys the expected constraint. For some of these benchmarks there are previous results for the special case in which the constraints are linear. The current paper improves over them both in terms of the generality of the constraint, and in terms of the quality of the guarantees.
Strengths: The paper considers a very general and natural model, and improves in this model over the previous state-of-the-art. The paper also presents a single algorithm that handles both adversarial and stochastic regrets. However, I believe that the last point is less impressive than what one might expect due to the following reason. Algorithms in submodular maximization tend to be quite natural and simple, and the sophistication is mostly in their analyses. Since the algorithm of the paper is of that kind, the fact that the same natural strategy works with respect to different benchmarks (with different analyses) is not very surprising.
Weaknesses: The paper seems to be a bit incremental in terms of the techniques. The algorithms presented are quite similar to previous algorithms for related problems. There might be more novelty in their analysis, but this is not discussed at all in the main part of the paper.
Correctness: Unfortunately, I am unable to assess the correctness of the paper. All the proofs have been deferred to the additional material, which I have only skimmed. However, the results look plausible.
Clarity: Most the paper is well written. In particular, the paper tries to convey in a nice way the intuition behind its algorithms despite the fact that naturally the proofs are deferred to the additional material. The exception is the experiments section, which is very dense, and almost decipherable without first reading about the studied application in the additional material.
Relation to Prior Work: Yes, the previous work is extensively discussed.
Additional Feedback: - The response was read, and answers my questions regarding the horizon and the choice of the objective functions. Thus, I keep my (positive) score. - I did not like the deferring of the notation to the appendix. This makes it basically impossible to read the main paper without the appendix, which contradicts the original reason for the separation between the two. - One should explain already in the introduction that the horizon T is known beforehand. This might be well-known for people who are familiar with the particular previous works to which the current paper relates, but it is not obvious for others. - More information about the experiments should appear in the main part of the paper. In particular, I would like to see an explanation for the choice of the objective functions used.
Summary and Contributions: This paper investigates the solution of online optimization problems where the objective to maximize is DR-submodular and (a set of) convex averaged constraints must be satisfied. The paper proposes a novel algorithm, provides theoretical analysis and presents experimental results.
Strengths: The paper is extremely well written an easy to follow. The problem is relevant both theoretically and practically. The problem is timely and topically. Generalizes and unifies previous existing results.
Weaknesses: I am happy with the paper as it is. The format of some equations could be enhanced (size of parenthesis, inline fractions...), but this is up to the authors.
Correctness: Yes, they are.
Clarity: Extremely clear.
Relation to Prior Work: First round: Yes, it is. Additional comments/works in the context of stochastic dual optimization could have been included, but this is a minor issue. Second round: As I said in the first round, I think that the paper is a solid contribution. I understand that there are some similarities with respect to previous works (related algorithms) but the theoretical analysis here is stronger, so that I do not see a problem there. BTW, it is not clear if my suggestion will be taken into account (for sure, the authors' response does not explicitly mention it).
Additional Feedback: If possible change the format of some of the equations, write a couple of comments on the relation to stochastic dual optimization and expand the numerical results (in the supplementary material, suggested, not required).