Sleipnir C++ API
Loading...
Searching...
No Matches
problem.hpp
1// Copyright (c) Sleipnir contributors
2
3#pragma once
4
5#include <algorithm>
6#include <array>
7#include <cmath>
8#include <concepts>
9#include <functional>
10#include <iterator>
11#include <memory>
12#include <optional>
13#include <ranges>
14#include <utility>
15
16#include <Eigen/Core>
17#include <Eigen/SparseCore>
18#include <gch/small_vector.hpp>
19
20#include "sleipnir/autodiff/expression_type.hpp"
21#include "sleipnir/autodiff/gradient.hpp"
22#include "sleipnir/autodiff/hessian.hpp"
23#include "sleipnir/autodiff/jacobian.hpp"
24#include "sleipnir/autodiff/variable.hpp"
25#include "sleipnir/autodiff/variable_matrix.hpp"
26#include "sleipnir/optimization/solver/exit_status.hpp"
27#include "sleipnir/optimization/solver/interior_point.hpp"
28#include "sleipnir/optimization/solver/iteration_info.hpp"
29#include "sleipnir/optimization/solver/newton.hpp"
30#include "sleipnir/optimization/solver/options.hpp"
31#include "sleipnir/optimization/solver/sqp.hpp"
32#include "sleipnir/optimization/solver/util/bounds.hpp"
33#include "sleipnir/util/empty.hpp"
34#include "sleipnir/util/print.hpp"
35#include "sleipnir/util/print_diagnostics.hpp"
36#include "sleipnir/util/profiler.hpp"
37#include "sleipnir/util/spy.hpp"
38#include "sleipnir/util/symbol_exports.hpp"
39
40namespace slp {
41
65template <typename Scalar>
66class Problem {
67 public:
70
76 [[nodiscard]]
78 m_decision_variables.emplace_back();
79 return m_decision_variables.back();
80 }
81
89 [[nodiscard]]
90 VariableMatrix<Scalar> decision_variable(int rows, int cols = 1) {
91 m_decision_variables.reserve(m_decision_variables.size() + rows * cols);
92
93 VariableMatrix<Scalar> vars{detail::empty, rows, cols};
94
95 for (int row = 0; row < rows; ++row) {
96 for (int col = 0; col < cols; ++col) {
97 m_decision_variables.emplace_back();
98 vars[row, col] = m_decision_variables.back();
99 }
100 }
101
102 return vars;
103 }
104
116 [[nodiscard]]
118 // We only need to store the lower triangle of an n x n symmetric matrix;
119 // the other elements are duplicates. The lower triangle has (n² + n)/2
120 // elements.
121 //
122 // n
123 // Σ k = (n² + n)/2
124 // k=1
125 m_decision_variables.reserve(m_decision_variables.size() +
126 (rows * rows + rows) / 2);
127
128 VariableMatrix<Scalar> vars{detail::empty, rows, rows};
129
130 for (int row = 0; row < rows; ++row) {
131 for (int col = 0; col <= row; ++col) {
132 m_decision_variables.emplace_back();
133 vars[row, col] = m_decision_variables.back();
134 vars[col, row] = m_decision_variables.back();
135 }
136 }
137
138 return vars;
139 }
140
150 void minimize(const Variable<Scalar>& cost) { m_f = cost; }
151
161 void minimize(Variable<Scalar>&& cost) { m_f = std::move(cost); }
162
173 // Maximizing a cost function is the same as minimizing its negative
174 m_f = -objective;
175 }
176
187 // Maximizing a cost function is the same as minimizing its negative
188 m_f = -std::move(objective);
189 }
190
196 m_equality_constraints.reserve(m_equality_constraints.size() +
197 constraint.constraints.size());
198 std::ranges::copy(constraint.constraints,
199 std::back_inserter(m_equality_constraints));
200 }
201
207 m_equality_constraints.reserve(m_equality_constraints.size() +
208 constraint.constraints.size());
209 std::ranges::copy(constraint.constraints,
210 std::back_inserter(m_equality_constraints));
211 }
212
218 m_inequality_constraints.reserve(m_inequality_constraints.size() +
219 constraint.constraints.size());
220 std::ranges::copy(constraint.constraints,
221 std::back_inserter(m_inequality_constraints));
222 }
223
229 m_inequality_constraints.reserve(m_inequality_constraints.size() +
230 constraint.constraints.size());
231 std::ranges::copy(constraint.constraints,
232 std::back_inserter(m_inequality_constraints));
233 }
234
238 ExpressionType cost_function_type() const {
239 if (m_f) {
240 return m_f.value().type();
241 } else {
242 return ExpressionType::NONE;
243 }
244 }
245
249 ExpressionType equality_constraint_type() const {
250 if (!m_equality_constraints.empty()) {
251 return std::ranges::max(m_equality_constraints, {},
253 .type();
254 } else {
255 return ExpressionType::NONE;
256 }
257 }
258
262 ExpressionType inequality_constraint_type() const {
263 if (!m_inequality_constraints.empty()) {
264 return std::ranges::max(m_inequality_constraints, {},
266 .type();
267 } else {
268 return ExpressionType::NONE;
269 }
270 }
271
280 ExitStatus solve(const Options& options = Options{},
281 [[maybe_unused]] bool spy = false) {
282 using DenseVector = Eigen::Vector<Scalar, Eigen::Dynamic>;
283 using SparseMatrix = Eigen::SparseMatrix<Scalar>;
284 using SparseVector = Eigen::SparseVector<Scalar>;
285
286 // Create the initial value column vector
287 DenseVector x{m_decision_variables.size()};
288 for (size_t i = 0; i < m_decision_variables.size(); ++i) {
289 x[i] = m_decision_variables[i].value();
290 }
291
292 if (options.diagnostics) {
293 print_exit_conditions(options);
294 print_problem_analysis();
295 }
296
297 // Get the highest order constraint expression types
298 auto f_type = cost_function_type();
299 auto c_e_type = equality_constraint_type();
300 auto c_i_type = inequality_constraint_type();
301
302 // If the problem is empty or constant, there's nothing to do
303 if (f_type <= ExpressionType::CONSTANT &&
304 c_e_type <= ExpressionType::CONSTANT &&
305 c_i_type <= ExpressionType::CONSTANT) {
306#ifndef SLEIPNIR_DISABLE_DIAGNOSTICS
307 if (options.diagnostics) {
308 slp::println("\nInvoking no-op solver...\n");
309 }
310#endif
311 return ExitStatus::SUCCESS;
312 }
313
314 gch::small_vector<SetupProfiler> ad_setup_profilers;
315 ad_setup_profilers.emplace_back("setup").start();
316
317 VariableMatrix<Scalar> x_ad{m_decision_variables};
318
319 // Set up cost function
320 Variable f = m_f.value_or(Scalar(0));
321
322 // Set up gradient autodiff
323 ad_setup_profilers.emplace_back(" ↳ ∇f(x)").start();
324 Gradient g{f, x_ad};
325 ad_setup_profilers.back().stop();
326
327 int num_decision_variables = m_decision_variables.size();
328 int num_equality_constraints = m_equality_constraints.size();
329 int num_inequality_constraints = m_inequality_constraints.size();
330
331 gch::small_vector<std::function<bool(const IterationInfo<Scalar>& info)>>
332 iteration_callbacks;
333 for (const auto& callback : m_iteration_callbacks) {
334 iteration_callbacks.emplace_back(callback);
335 }
336 for (const auto& callback : m_persistent_iteration_callbacks) {
337 iteration_callbacks.emplace_back(callback);
338 }
339
340 // Solve the optimization problem
341 ExitStatus status;
342 if (m_equality_constraints.empty() && m_inequality_constraints.empty()) {
343 if (options.diagnostics) {
344 slp::println("\nInvoking Newton solver...\n");
345 }
346
347 // Set up Lagrangian Hessian autodiff
348 ad_setup_profilers.emplace_back(" ↳ ∇²ₓₓL").start();
349 Hessian<Scalar, Eigen::Lower> H{f, x_ad};
350 ad_setup_profilers.back().stop();
351
352 ad_setup_profilers[0].stop();
353
354#ifndef SLEIPNIR_DISABLE_DIAGNOSTICS
355 // Sparsity pattern files written when spy flag is set
356 std::unique_ptr<Spy<Scalar>> H_spy;
357
358 if (spy) {
359 H_spy = std::make_unique<Spy<Scalar>>(
360 "H.spy", "Hessian", "Decision variables", "Decision variables",
361 num_decision_variables, num_decision_variables);
362 iteration_callbacks.push_back(
363 [&](const IterationInfo<Scalar>& info) -> bool {
364 H_spy->add(info.H);
365 return false;
366 });
367 }
368#endif
369
370 NewtonMatrixCallbacks<Scalar> matrix_callbacks{
371 num_decision_variables,
372 [&](const DenseVector& x) -> Scalar {
373 x_ad.set_value(x);
374 return f.value();
375 },
376 [&](const DenseVector& x) -> SparseVector {
377 x_ad.set_value(x);
378 return g.value();
379 },
380 [&](const DenseVector& x) -> SparseMatrix {
381 x_ad.set_value(x);
382 return H.value();
383 }};
384
385 // Invoke Newton solver
386 status =
387 newton<Scalar>(matrix_callbacks, iteration_callbacks, options, x);
388 } else if (m_inequality_constraints.empty()) {
389 if (options.diagnostics) {
390 slp::println("\nInvoking SQP solver\n");
391 }
392
393 VariableMatrix<Scalar> c_e_ad{m_equality_constraints};
394
395 // Set up equality constraint Jacobian autodiff
396 ad_setup_profilers.emplace_back(" ↳ ∂cₑ/∂x").start();
397 Jacobian A_e{c_e_ad, x_ad};
398 ad_setup_profilers.back().stop();
399
400 // Set up Lagrangian
401 VariableMatrix<Scalar> y_ad(num_equality_constraints);
402 Variable L = f - y_ad.T() * c_e_ad;
403
404 // Set up Lagrangian Hessian autodiff
405 ad_setup_profilers.emplace_back(" ↳ ∇²ₓₓL").start();
406 Hessian<Scalar, Eigen::Lower> H{L, x_ad};
407 ad_setup_profilers.back().stop();
408
409 ad_setup_profilers[0].stop();
410
411#ifndef SLEIPNIR_DISABLE_DIAGNOSTICS
412 // Sparsity pattern files written when spy flag is set
413 std::unique_ptr<Spy<Scalar>> H_spy;
414 std::unique_ptr<Spy<Scalar>> A_e_spy;
415
416 if (spy) {
417 H_spy = std::make_unique<Spy<Scalar>>(
418 "H.spy", "Hessian", "Decision variables", "Decision variables",
419 num_decision_variables, num_decision_variables);
420 A_e_spy = std::make_unique<Spy<Scalar>>(
421 "A_e.spy", "Equality constraint Jacobian", "Constraints",
422 "Decision variables", num_equality_constraints,
423 num_decision_variables);
424 iteration_callbacks.push_back(
425 [&](const IterationInfo<Scalar>& info) -> bool {
426 H_spy->add(info.H);
427 A_e_spy->add(info.A_e);
428 return false;
429 });
430 }
431#endif
432
433 SQPMatrixCallbacks<Scalar> matrix_callbacks{
434 num_decision_variables,
435 num_equality_constraints,
436 [&](const DenseVector& x) -> Scalar {
437 x_ad.set_value(x);
438 return f.value();
439 },
440 [&](const DenseVector& x) -> SparseVector {
441 x_ad.set_value(x);
442 return g.value();
443 },
444 [&](const DenseVector& x, const DenseVector& y) -> SparseMatrix {
445 x_ad.set_value(x);
446 y_ad.set_value(y);
447 return H.value();
448 },
449 [&](const DenseVector& x) -> DenseVector {
450 x_ad.set_value(x);
451 return c_e_ad.value();
452 },
453 [&](const DenseVector& x) -> SparseMatrix {
454 x_ad.set_value(x);
455 return A_e.value();
456 }};
457
458 // Invoke SQP solver
459 status = sqp<Scalar>(matrix_callbacks, iteration_callbacks, options, x);
460 } else {
461 if (options.diagnostics) {
462 slp::println("\nInvoking IPM solver...\n");
463 }
464
465 VariableMatrix<Scalar> c_e_ad{m_equality_constraints};
466 VariableMatrix<Scalar> c_i_ad{m_inequality_constraints};
467
468 // Set up equality constraint Jacobian autodiff
469 ad_setup_profilers.emplace_back(" ↳ ∂cₑ/∂x").start();
470 Jacobian A_e{c_e_ad, x_ad};
471 ad_setup_profilers.back().stop();
472
473 // Set up inequality constraint Jacobian autodiff
474 ad_setup_profilers.emplace_back(" ↳ ∂cᵢ/∂x").start();
475 Jacobian A_i{c_i_ad, x_ad};
476 ad_setup_profilers.back().stop();
477
478 // Set up Lagrangian
479 VariableMatrix<Scalar> y_ad(num_equality_constraints);
480 VariableMatrix<Scalar> z_ad(num_inequality_constraints);
481 Variable L = f - y_ad.T() * c_e_ad - z_ad.T() * c_i_ad;
482
483 // Set up Lagrangian Hessian autodiff
484 ad_setup_profilers.emplace_back(" ↳ ∇²ₓₓL").start();
485 Hessian<Scalar, Eigen::Lower> H{L, x_ad};
486 ad_setup_profilers.back().stop();
487
488 ad_setup_profilers[0].stop();
489
490#ifndef SLEIPNIR_DISABLE_DIAGNOSTICS
491 // Sparsity pattern files written when spy flag is set
492 std::unique_ptr<Spy<Scalar>> H_spy;
493 std::unique_ptr<Spy<Scalar>> A_e_spy;
494 std::unique_ptr<Spy<Scalar>> A_i_spy;
495
496 if (spy) {
497 H_spy = std::make_unique<Spy<Scalar>>(
498 "H.spy", "Hessian", "Decision variables", "Decision variables",
499 num_decision_variables, num_decision_variables);
500 A_e_spy = std::make_unique<Spy<Scalar>>(
501 "A_e.spy", "Equality constraint Jacobian", "Constraints",
502 "Decision variables", num_equality_constraints,
503 num_decision_variables);
504 A_i_spy = std::make_unique<Spy<Scalar>>(
505 "A_i.spy", "Inequality constraint Jacobian", "Constraints",
506 "Decision variables", num_inequality_constraints,
507 num_decision_variables);
508 iteration_callbacks.push_back(
509 [&](const IterationInfo<Scalar>& info) -> bool {
510 H_spy->add(info.H);
511 A_e_spy->add(info.A_e);
512 A_i_spy->add(info.A_i);
513 return false;
514 });
515 }
516#endif
517
518 const auto [bound_constraint_mask, bounds, conflicting_bound_indices] =
519 get_bounds<Scalar>(m_decision_variables, m_inequality_constraints,
520 A_i.value());
521 if (!conflicting_bound_indices.empty()) {
522 if (options.diagnostics) {
523 print_bound_constraint_global_infeasibility_error(
524 conflicting_bound_indices);
525 }
526 return ExitStatus::GLOBALLY_INFEASIBLE;
527 }
528
529#ifdef SLEIPNIR_ENABLE_BOUND_PROJECTION
530 project_onto_bounds(x, bounds);
531#endif
532
533 InteriorPointMatrixCallbacks<Scalar> matrix_callbacks{
534 num_decision_variables,
535 num_equality_constraints,
536 num_inequality_constraints,
537 [&](const DenseVector& x) -> Scalar {
538 x_ad.set_value(x);
539 return f.value();
540 },
541 [&](const DenseVector& x) -> SparseVector {
542 x_ad.set_value(x);
543 return g.value();
544 },
545 [&](const DenseVector& x, const DenseVector& y,
546 const DenseVector& z) -> SparseMatrix {
547 x_ad.set_value(x);
548 y_ad.set_value(y);
549 z_ad.set_value(z);
550 return H.value();
551 },
552 [&](const DenseVector& x) -> DenseVector {
553 x_ad.set_value(x);
554 return c_e_ad.value();
555 },
556 [&](const DenseVector& x) -> SparseMatrix {
557 x_ad.set_value(x);
558 return A_e.value();
559 },
560 [&](const DenseVector& x) -> DenseVector {
561 x_ad.set_value(x);
562 return c_i_ad.value();
563 },
564 [&](const DenseVector& x) -> SparseMatrix {
565 x_ad.set_value(x);
566 return A_i.value();
567 }};
568
569 // Invoke interior-point method solver
570 status =
571 interior_point<Scalar>(matrix_callbacks, iteration_callbacks, options,
572#ifdef SLEIPNIR_ENABLE_BOUND_PROJECTION
573 bound_constraint_mask,
574#endif
575 x);
576 }
577
578 if (options.diagnostics) {
579 print_autodiff_diagnostics(ad_setup_profilers);
580 slp::println("\nExit: {}", status);
581 }
582
583 // Assign the solution to the original Variable instances
584 VariableMatrix<Scalar>{m_decision_variables}.set_value(x);
585
586 return status;
587 }
588
594 template <typename F>
595 requires requires(F callback, const IterationInfo<Scalar>& info) {
596 { callback(info) } -> std::same_as<void>;
597 }
599 m_iteration_callbacks.emplace_back(
600 [=, callback =
601 std::forward<F>(callback)](const IterationInfo<Scalar>& info) {
602 callback(info);
603 return false;
604 });
605 }
606
613 template <typename F>
614 requires requires(F callback, const IterationInfo<Scalar>& info) {
615 { callback(info) } -> std::same_as<bool>;
616 }
618 m_iteration_callbacks.emplace_back(std::forward<F>(callback));
619 }
620
622 void clear_callbacks() { m_iteration_callbacks.clear(); }
623
632 template <typename F>
633 requires requires(F callback, const IterationInfo<Scalar>& info) {
634 { callback(info) } -> std::same_as<bool>;
635 }
637 m_persistent_iteration_callbacks.emplace_back(std::forward<F>(callback));
638 }
639
640 private:
641 // The list of decision variables, which are the root of the problem's
642 // expression tree
643 gch::small_vector<Variable<Scalar>> m_decision_variables;
644
645 // The cost function: f(x)
646 std::optional<Variable<Scalar>> m_f;
647
648 // The list of equality constraints: cₑ(x) = 0
649 gch::small_vector<Variable<Scalar>> m_equality_constraints;
650
651 // The list of inequality constraints: cᵢ(x) ≥ 0
652 gch::small_vector<Variable<Scalar>> m_inequality_constraints;
653
654 // The iteration callbacks
655 gch::small_vector<std::function<bool(const IterationInfo<Scalar>& info)>>
656 m_iteration_callbacks;
657 gch::small_vector<std::function<bool(const IterationInfo<Scalar>& info)>>
658 m_persistent_iteration_callbacks;
659
660 void print_exit_conditions([[maybe_unused]] const Options& options) {
661 // Print possible exit conditions
662 slp::println("User-configured exit conditions:");
663 slp::println(" ↳ error below {}", options.tolerance);
664 if (!m_iteration_callbacks.empty() ||
665 !m_persistent_iteration_callbacks.empty()) {
666 slp::println(" ↳ iteration callback requested stop");
667 }
668 if (std::isfinite(options.max_iterations)) {
669 slp::println(" ↳ executed {} iterations", options.max_iterations);
670 }
671 if (std::isfinite(options.timeout.count())) {
672 slp::println(" ↳ {} elapsed", options.timeout);
673 }
674 }
675
676 void print_problem_analysis() {
677 constexpr std::array types{"no", "constant", "linear", "quadratic",
678 "nonlinear"};
679
680 // Print problem structure
681 slp::println("\nProblem structure:");
682 slp::println(" ↳ {} cost function",
683 types[std::to_underlying(cost_function_type())]);
684 slp::println(" ↳ {} equality constraints",
685 types[std::to_underlying(equality_constraint_type())]);
686 slp::println(" ↳ {} inequality constraints",
687 types[std::to_underlying(inequality_constraint_type())]);
688
689 if (m_decision_variables.size() == 1) {
690 slp::print("\n1 decision variable\n");
691 } else {
692 slp::print("\n{} decision variables\n", m_decision_variables.size());
693 }
694
695 auto print_constraint_types =
696 [](const gch::small_vector<Variable<Scalar>>& constraints) {
697 std::array<size_t, 5> counts{};
698 for (const auto& constraint : constraints) {
699 ++counts[std::to_underlying(constraint.type())];
700 }
701 for (const auto& [count, name] :
702 std::views::zip(counts, std::array{"empty", "constant", "linear",
703 "quadratic", "nonlinear"})) {
704 if (count > 0) {
705 slp::println(" ↳ {} {}", count, name);
706 }
707 }
708 };
709
710 // Print constraint types
711 if (m_equality_constraints.size() == 1) {
712 slp::println("1 equality constraint");
713 } else {
714 slp::println("{} equality constraints", m_equality_constraints.size());
715 }
716 print_constraint_types(m_equality_constraints);
717 if (m_inequality_constraints.size() == 1) {
718 slp::println("1 inequality constraint");
719 } else {
720 slp::println("{} inequality constraints",
721 m_inequality_constraints.size());
722 }
723 print_constraint_types(m_inequality_constraints);
724 }
725};
726
727extern template class EXPORT_TEMPLATE_DECLARE(SLEIPNIR_DLLEXPORT)
728Problem<double>;
729
730} // namespace slp
Definition intrusive_shared_ptr.hpp:27
Definition problem.hpp:66
VariableMatrix< Scalar > symmetric_decision_variable(int rows)
Definition problem.hpp:117
void subject_to(InequalityConstraints< Scalar > &&constraint)
Definition problem.hpp:228
void add_callback(F &&callback)
Definition problem.hpp:617
void subject_to(EqualityConstraints< Scalar > &&constraint)
Definition problem.hpp:206
ExpressionType equality_constraint_type() const
Definition problem.hpp:249
void maximize(Variable< Scalar > &&objective)
Definition problem.hpp:186
void add_persistent_callback(F &&callback)
Definition problem.hpp:636
void subject_to(const InequalityConstraints< Scalar > &constraint)
Definition problem.hpp:217
ExpressionType inequality_constraint_type() const
Definition problem.hpp:262
void minimize(const Variable< Scalar > &cost)
Definition problem.hpp:150
Problem() noexcept=default
Construct the optimization problem.
VariableMatrix< Scalar > decision_variable(int rows, int cols=1)
Definition problem.hpp:90
void minimize(Variable< Scalar > &&cost)
Definition problem.hpp:161
void clear_callbacks()
Clears the registered callbacks.
Definition problem.hpp:622
void subject_to(const EqualityConstraints< Scalar > &constraint)
Definition problem.hpp:195
void maximize(const Variable< Scalar > &objective)
Definition problem.hpp:172
void add_callback(F &&callback)
Definition problem.hpp:598
ExpressionType cost_function_type() const
Definition problem.hpp:238
ExitStatus solve(const Options &options=Options{}, bool spy=false)
Definition problem.hpp:280
Variable< Scalar > decision_variable()
Definition problem.hpp:77
Definition variable.hpp:47
Solver options.
Definition options.hpp:13