Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Blockchain, 03 December 2025

Sec. Blockchain Economics

Volume 8 - 2025 | https://doi.org/10.3389/fbloc.2025.1648418

This article is part of the Research TopicBlockchain and Tokenomics for Sustainable Development, Volume IIView all 3 articles

Thermodynamic control of patent tokenization for sustainable development

  • University of Applied Sciences Rosenheim, Rosenheim, Germany

Background: Patent tokenization converts intellectual property (IP) into tradable digital units. Pilots on IPwe, IBM, and Ocean Protocol have processed >1,500 assets ($62 million since 2019). In practice, schedules are often fixed and pricing simplified; risk preferences are typically ignored—costing value in volatile settings relevant to the SDGs.

Methods: We cast patent tokenization as risk-sensitive control and linked economic risk aversion to inverse temperature. With the exponential transform Φ=eγV, the risk-sensitive Hamilton–Jacobi–Bellman (HJB) linearizes for fixed controls; the optimal policy is a pointwise threshold (bang–bang or smoothly regularized). This preserves tractability and enables millisecond solves.

Results: From 45 tokenization trajectories (12 held out), we estimate γ=2.1±0.38 (95% CI), while a composite objective F(γ) calibrated to equity-like orders peaks at γ*3.02 ([2.9,3.2]). Monte Carlo (n=10,000) shows a 76% downside reduction and 92% success rate under ledger-based schedules. Solver latency drops by 850× (40 s 47 ms).

Conclusion: The thermodynamic lens explains why practice (γ2.1) deviates from the theoretical benchmark (γ*3.0) in our calibration: actors appear to trade maximal efficiency for robustness. The framework stabilizes revenues, drives fees to negligible levels, and supports SDG-aligned innovation finance by lowering the cost of access and adaptation under uncertainty.

1 Introduction

1.1 Orientation

Patent tokenization aims to unlock illiquid IP under uncertainty. Static release plans and simplified pricing cannot adapt to shocks (litigation, regulation, and demand), and they rarely encode risk preferences, which is precisely where value is at stake in SDG-relevant domains.

1.2 This paper’s idea in one line

We treat patent tokenization as a risk-sensitive control and use a thermodynamic change of variables to make the problem linear for fixed controls so that optimal schedules follow from a simple threshold structure and can be computed in milliseconds.

1.3 Contributions

Theory. The risk-sensitive HJB with the correct sign convention linearizes under Φ=eγV and implies threshold (bang–bang) policies; a small quadratic penalty yields smooth, implementable schedules.

Evidence. From 45 trajectories, we estimate γ=2.1±0.38, while a composite objective peaks at γ*3.02 under our calibration. Monte Carlo (n=10,000) indicates 76% downside reduction, 92% success, and 850× speedup.

SDG context. The framework clarifies when robust, risk-aware schedules are preferable to fixed ones, lowering access costs and supporting SDG-aligned innovation (e.g., infrastructure and climate-related technologies).

1.4 Roadmap

Section 3 derives the transform and control structure. Section 4 reports estimates and simulations, and Section 5 interprets the robustness–efficiency trade-off and SDG implications; the Supplementary Information documents sensitivity and implementation details.

The global patent system contains vast reserves of unused value (World Intellectual Property Organization WIPO, 2024; European Patent Office EPO, 2023; United States Patent and Trademark Office USPTO, 2024). Although patents are designed to secure innovation, empirical studies show that between 90% and 95% never generate licensing income (Gambardella et al., 2007; Arora et al., 2008).

One reason is the rigid all-or-nothing structure of conventional transactions: patent holders either retain their rights completely, leaving them without immediate funding, or sell them outright and forgo future upsides (Lemley and Shapiro, 2005; Lanjouw and Schankerman, 2004; Bessen and Meurer, 2008; Boldrin and Levine, 2008; Scotchmer, 2004). Distributed ledger technology offers an alternative by enabling fractional and tradable ownership through tokenization. Digital rights can be created and exchanged on blockchain networks, opening new channels for liquidity (Catalini and Gans, 2018; Howell et al., 2020; Benedetti and Kostovetsky, 2021; Chen, 2018; Cong and He, 2019; Yermack, 2017; Malinova and Park, 2017). Nevertheless, deciding when and how to release tokens under uncertainty remains difficult since poorly chosen schedules can materially erode value (Akcigit and Kerr, 2018; Bloom et al., 2020; Jones and Williams, 2000; Kortum and Lerner, 2000).

1.5 Research topic context

This article is aligned with the Frontiers research topic Blockchain and Tokenomics for Sustainable Development (Volume II) and follows its call for sustainability-driven tokenomics and governance models (Frontiers, 2025).

Practical experiments have already demonstrated feasibility. IPwe tokenized more than 500 patents using Black–Scholes pricing and recorded approximately 25 million USD in transaction volume. IBM placed over 800 patents on Hyperledger with fixed release plans and reached roughly 31 million USD. WIPO ran a pilot with more than 200 patents. Ocean Protocol applied bonding curves to intellectual property and data, generating close to 6 million USD (IPwe, 2024; IBM, 2022; Deloitte, 2023).

These cases show what is possible but also highlight structural weaknesses. Existing platforms do not incorporate risk preferences, they rely on rigid schedules that cannot adapt to shocks such as litigation, and they charge fees in the percent range instead of technically attainable near-negligible levels. Their dynamics follow known patterns of multi-sided platforms (Hagiu and Wright, 2015; Rochet and Tirole, 2006; Evans and Schmalensee, 2008; Rysman, 2009).

This study embeds risk-aware dynamic optimization into tokenization using a thermodynamic perspective. The key step is transforming the nonlinear Hamilton–Jacobi–Bellman equation into a linear partial differential equation via Φ=eγV, exploiting the negative sign of the quadratic gradient term from risk-sensitive control (Fleming and Soner, 2006). Consistent with path-integral control (Kappen, 2005; Todorov, 2009), this turns an otherwise intractable problem into one solvable in milliseconds. In data from 45 tokenized assets (real estate, art, and patents), the empirical estimate of risk aversion is γ=2.1±0.38, while a composite objective calibrated to market-like parameters places the optimum near γ*3.0.

Simulations confirm the gap: the empirical value yields the highest survival rate, and the theoretical benchmark maximizes efficiency. Beyond the immediate application, links to the free-energy principle (Friston, 2010), maximum entropy (Jaynes, 1957), and large deviations (Touchette, 2009) point to deeper structural parallels between economic decision-making and adaptive systems. Together, these elements show how a thermodynamic approach can make tokenization more flexible, risk-sensitive, and efficient.

2 Problem statement

Patent tokenization is a continuous-time stochastic control task (Fleming and Soner, 2006; Øksendal, 2003). Notation is established early for clarity:

Xt: monetization basis (current market value of patent) following geometric Brownian motion (Øksendal, 2003; Campbell et al., 1997);

ft: remaining fraction of patent ownership (starts at 1 and decreases to 0);

ut: tokenization release rate (control variable, units: 1/time); and

V(t,x,f): value function representing expected future revenue.

We use uppercase Xt for the stochastic process and lowercase x for the PDE state variable and the same for ft versus f. For a risk-aware decision-maker with a constant absolute risk-aversion (CARA) level γ, this results in the Hamilton-Jacobi-Bellman (HJB) equation (Equation 2.1; Fleming and Soner, 2006):

Vt=maxu0,umaxxu+μxVxuVf+σ2x222Vx2γσ2x22Vx2,(2.1)

where (x,u)=xu is the immediate revenue from tokenization, and L=μxx+σ2x222x2 is the infinitesimal generator of geometric Brownian motion (Øksendal, 2003). The boundary conditions are

V(T,x,f)=0 (no future value at terminal time) and

V(t,x,0)=0 (no remaining tokens means no future revenue).

Additionally, a rate limit 0utumax applies.

In the simulations, we use umax=1.0 per time unit, which allows full tokenization (from f=1 to f=0) by T=1. For the discrete implementation, we use a grid in f with Δf=1/(Nf1) and choose Δt=Δf/umax. Hence, the one time-step reduces f by exactly one grid cell under umax, and full tokenization is feasible within T=1. This sign is essential for the thermodynamic transformation and follows from risk-sensitive control theory (Fleming and Soner, 2006).

3 Methods

3.1 Thermodynamic transformation

Our analysis begins with the risk-sensitive Hamilton–Jacobi–Bellman (HJB) equation. To facilitate readability, we reproduce it here once more in the version adapted to our framework:

Vt=maxu0,umaxxu+μxVxuVf+σ2x222Vx2γσ2x22Vx2.(3.1)

It is critical to note the negative sign in front of the quadratic gradient term. Without this sign, the thermodynamic transformation would fail because the nonlinear contribution would survive. Based on this structure, as shown in risk-sensitive control theory (Fleming and Soner, 2006), the transformation becomes straightforward. We define the substitution,

Φt,x,f=expγVt,x,f.

By direct differentiation, we obtain the relations collected in Equation 3.2-3.6.

Φt=γΦVt,(3.2)
Φx=γΦVx,(3.3)
Vx=1γΦΦx,(3.4)
Φf=γΦVfVf=1γΦΦf,(3.5)
2Φx2=γΦ2Vx2+γ2ΦVx2.(3.6)

One key observation is that the nonlinear terms cancel exactly when substituted into Equation 3.1. The second derivative contributes a positive quadratic term, which is offset by the negative quadratic term already present in the HJB. The algebra looks heavy at first, but the cancellation is straightforward once written out. What matters in practice is that the nonlinear term drops out and we end up with a linear PDE. This makes the difference between a system that is computationally intractable and one that can be solved quickly enough for actual tokenization platforms:

Φt+μxΦx+σ2x222Φx2=maxu0,umaxγxuΦ+uΦf.(3.7)

3.1.1 Important clarification

For any fixed control u, this PDE is linear in Φ. The overall control problem remains a threshold-based optimization due to the pointwise maximization over u[0,umax], but the computational advantages of linearity are preserved.

3.1.2 Discrete-time Bellman recursion in Φ-space (implementation)

With Δt chosen such that Δf=umaxΔt, the numerically implemented risk-sensitive dynamic program in Φ-space (with quadratic smoothness penalty κ2u2 inside the reward ) reads

Φt,x,f=minu0,umaxEexpγxuκ2u2ΔtΦt+Δt,X,fuΔt,

where X=xexp(μ12σ2)Δt+σΔtZ with ZN(0,1). Expanding to first order in Δt recovers the linear PDE in Equation 3.7 (with a pointwise maximization in V-space equivalent to the minimization above in Φ-space).

3.2 Optimal control structure

The optimal control emerges from maximizing the Hamiltonian in Equation 3.8.

Hu=xuuVf.(3.8)

Since H is linear in u (not quadratic), we have 2Hu2=0. Therefore, the control structure is not determined by phase transitions but by the sign of Hu=xVf combined with the box constraints:

u*=umax,if x>Vfbang-bang high,0,if x<Vfbang-bang low,any u0,umax,if x=Vfsingular.(3.9)

If we add a small penalty κ2u2 to the cost function, we obtain H(u)=xuVfuκ2u2 and thus the interior point u(t,x,f)=Π[0,umax]((xVf)/κ). For κ0, the resulting bang-bang structure shown in Equation 3.9 for small κ>0, smooth controls arise. The apparent smoothness in certain parameter regimes arises from

numerical regularization/discretization in implementation;

transaction costs and rate limits in practice; and

explicit penalty terms added for computational stability (e.g., κ2u2 term).

These are implementation choices or practical frictions, not intrinsic phase transitions of the Hamiltonian itself.

3.2.1 Smoothness via explicit quadratic penalty

For the figures, we include a small quadratic penalty κ2u2 with κ=0.02. This changes the Hamiltonian to H(u)=xuVfuκ2u2, which admits the interior optimizer

u=Π0,umaxxVfκ.

As κ0, the policy reduces to the bang–bang structure. With κ=0.02, the token release trajectories become smooth, consistent with Figures 14.

Figure 1
Graph titled

Figure 1. Tokenization trajectories for different risk preferences. Curves are generated with κ=0.02, which yields smooth release profiles. Risk-neutral (γ=0) delays release, empirical estimate (γ=2.1) follows a gradual S-shape, and the mathematical optimum (γ*3.02) is more front-loaded. Time normalized to t/T with umax=1.0. Smoothness arises from the quadratic control penalty (κ=0.02).

Figure 2
Graph showing revenue distributions for 10,000 Monte Carlo samples. The x-axis shows normalized total revenue and the y-axis shows probability density. Three curves are displayed: a red dashed curve for γ = 0 (risk-neutral), a solid blue curve for γ = 2.1 (empirical estimate), and a dotted curve for γ* ≈ 3.02 (optimum under the calibration). A pink shaded region labeled ‘Downside’ highlights revenue values below 0.2, indicating the downside-risk zone.

Figure 2. Revenue distributions from Monte Carlo simulation. Risk-neutral (γ=0) shows bimodal distribution with high variance. Empirical estimate from our sample (γ=2.1) concentrates around a mean revenue of 0.224. Mathematical optimum under our calibration (γ*3.02) has the narrowest distribution with mean 0.167. Shaded region indicates downside risk zone. Total revenue is normalized to initial value (dimensionless); probability density has units of inverse revenue. Additional robustness checks and extended versions are provided in the Supplementary Information (Supplementary Figures S1–S3). All distributions computed with κ=0.02, which regularizes bang–bang strategies into smooth tokenization. Smoothness arises from the quadratic control penalty (κ=0.02).

Figure 3
Graph titled

Figure 3. Empirical validation with real tokenization data. Gray markers show 37 real estate tokenization trajectories; γ=2.1±0.38 is from the full multi-asset dataset and does not imply that patents lie at γ=2.1. The blue curve with 95% bands shows the model at γ=2.1; the violet dashed curve shows the optimum at γ*3.02 under our calibration. Time is normalized to t/T (dimensionless); fraction tokenized is dimensionless (0–1). Additional robustness checks and extended versions are provided in the Supplementary Information (Supplementary Figures S1–S3). Smoothness arises from the quadratic control penalty (κ=0.02).

Figure 4
Line graph comparing the fraction tokenized over time for different models. The x-axis represents time \(t/T\), and the y-axis shows the fraction tokenized. There are four lines: IPwe with a fixed schedule (red dashed), IBM static (blue dotted), and two versions of the model with different \(\gamma\) values (solid blue and green dashed). An inset box indicates volatility reduction of forty-five percent, revenue stability increase of twenty-eight percent, and a success rate of ninety-two percent for twelve held-out assets.

Figure 4. Comparison of tokenization trajectories from real platforms versus our model. Curves labeled “IPwe” and “IBM” are indicative schedules based on publicly available platform documentation. Our model curves at γ=2.1 (empirical estimate from our sample) and γ*3.02 (optimum under our calibration) illustrate risk-sensitive schedules. The shaded region represents the 95% confidence band from bootstrap analysis. Time is normalized to t/T (dimensionless); fraction tokenized is dimensionless (0–1). Computational performance: 850× speedup (40 s 47 ms on NVIDIA RTX 3090). Additional robustness checks and extended versions are provided in the Supplementary Information (Supplementary Figures S1–S3). Smoothness arises from the quadratic control penalty (κ=0.02).

3.2.2 Threshold in V- vs. Φ-space

Since Φ=eγV, it holds Φf=γΦVf; hence,

γxΦ+Φf=γΦxVf.

The bang–bang rule u*=umax1{x>Vf} in V-space is therefore equivalent to u*=umax1{γxΦ+Φf>0} in Φ-space, consistent with Equation 3.7.

3.2.3 Max. in V-space vs. min. in Φ-space

Because Φ=eγV strictly decreases in V, the pointwise maximization in the HJB for V is equivalent to a minimization in the dynamic program for Φ. Both formulations are used: Equation 3.7 shows the V-space maximization mapped to Φ, while the implementable recursion is written as a minimization over u.

3.3 Multi-objective optimization framework

To determine the optimal risk parameter γ*, we define the multi-objective function F(γ) in Equation 3.10 that balances three critical performance dimensions of a tokenization strategy. The optimal parameter γ* is found by maximizing the function

γ*=argmaxγ>0Fγ=argmaxγ>0SγKγBγ.(3.10)

The components are defined as follows. These functional forms are modeling choices inspired by, but not identical to, standard financial metrics.

The modified Sharpe component S(γ) is given in Equation 3.11. This term is inspired by the Sharpe ratio (Sharpe, 1994; Markowitz, 1952) but adapted for our context. It is estimated through Monte Carlo simulation (Metropolis and Ulam, 1949; Glasserman, 2003; Boyle et al., 1997) with 10,000 paths:

Sγ=ERu*γrStd Ru*γ,R0TXtutdt,r=0.02,(3.11)

where r=0.02 is the risk-free return based on long-term averages for equity-like assets (Campbell et al., 1997; Fama and French, 2002). Note that when μ=r, the derivative S/γ vanishes, making all γ equally optimal under this criterion. Our calibration assumes μ>r. For tables and figures, we use an analytical surrogate for S(γ) that closely approximates the Monte Carlo estimate; however, the main evaluations are based on Monte Carlo with 10,000 paths. We evaluate S(γ) on simulated revenue R rather than on terminal asset value XT, since the control objective is tokenization revenue. For all simulations and code, we set the reference rate to r=0.02.

Kelly-distance component K(γ). The component K(γ) captures proximity to Kelly-like efficiency. While the classical Kelly fraction f*=(μr)/σ2 arises under log-utility (Merton, 1972), our CARA framework motivates a saturating proxy K(γ)=1exp(1.5γ). This formulation is inspired by the Kelly criterion but adapted to exponential utility.

The behavioral acceptance function B(γ) is defined in Equation 3.12. This term measures the alignment between the mathematical model and observed human behavior. Behavioral evidence places risk-aversion parameters for gains at approximately 2.0–2.25 (Kahneman and Tversky, 1979; Barberis, 2013), which motivates our calibration anchor:

Bγ=11+exp2γγb,(3.12)

where γb=2.0 is a calibration anchor based on behavioral finance literature, not a universal constant.

The product form SKB ensures that a strategy is only considered optimal if it performs well across all three dimensions simultaneously. This value is calibration-specific. Different plausible parameter choices (μ, σ, r) would shift γ*. Calibrations use equity-like orders of magnitude (Campbell et al., 1997; Fama and French, 2002). Sensitivity over (μ,σ,r) grids is reported in Supplementary Information section B.1.

4 Results

The previous section introduced the composite objective F(γ). We now apply this framework to our baseline calibration (μ,σ,r) and dataset. In that sense, we evaluate the framework on tokenization trajectories collected from blockchain platforms between 2019 and 2024. The dataset includes real estate (RealT, 2024; Baum, 2021), artworks (Maecenas, 2024; Kräussl and Whitaker, 2020), and patent implementations (IPwe, 2024; IBM, 2022; Deloitte, 2023). All of these assets share key properties with patents: they have uncertain and often volatile future value, limited liquidity in traditional markets, and they lend themselves naturally to fractional ownership structures (Hagiu and Wright, 2015; Rochet and Tirole, 2006; Evans and Schmalensee, 2008; Rysman, 2009). Estimation using maximum likelihood on 45 tokenization trajectories yields an empirical risk parameter of γ=2.1±0.38 (95% confidence interval) in our sample. This estimate captures how participants in the sample balance upside and downside when issuing tokens and reflects behavioral factors specific to our dataset. We considered the multi-objective function

Fγ=SγKγBγ,(4.1)

(Equation 4.1), as defined in Section 3.

Remark 1 (calibration-specific numerical finding). Under the calibration μ=0.05, σ=0.2, r=0.02, λ=1.5 and γb=2.0, the composite objective attains at γ*3.02 the value F(3.02)=0.0726576. At the empirical estimate γ=2.1, we obtain F(2.1)=0.0542061.

Note. This result is calibration- and dataset-specific, not a universal theorem.

Numerical evidence. With S(3.02)=0.083 and S(2.1)=0.103 from simulation, and closed forms K(γ)=1e1.5γ and B(γ)=1/(1+e2(γ2)), we obtain K(3.02)=0.9892193 and B(3.02)=0.8849333, hence F(3.02)=0.0726576, and K(2.1)=0.9571479 and B(2.1)=0.5498340; hence, F(2.1)=0.0542061. A finite-difference second-derivative test confirms a local maximum near γ*3.02.

The exact relative gap is

|F3.02F2.1|F3.02=0.253952225.395%.

Key result. Under our calibration, the maximum is at γ*3.02 with F=0.0726576, while the empirical estimator γ=2.1±0.38 yields F=0.0542061 (relative gap 25.395%).

4.1 Multi-agent simulation results

To understand the deviation in our sample, we simulated ten heterogeneous agents with risk-aversion levels ranging from γ=0.5 to γ=4.5. Each agent executed 2,000 Monte Carlo paths, giving 20,000 trajectories in total. The outcomes fall into three distinct behavioral regimes (not phase transitions). For γ<1.5, agents follow predominantly bang–bang strategies that swing between full tokenization and complete restraint. Between γ=1.5 and γ=4.5, token release becomes smoother due to numerical regularization and implementation choices. At the highest values, above γ=4.5, strategies become very conservative, with minimal tokenization. Performance metrics highlight these differences. The detailed results are summarized in Table 1. Agent 1 (γ=0.5) earns the largest average revenue but suffers drawdowns above 60% and a success rate of only 43%. Agent 5 (γ=2.1) achieves strong survival outcomes, with 92% of trajectories meeting the revenue threshold. Agent 7 (γ3.02) approaches the multi-objective optimum under our calibration, although its success rate is 89%. Finally, Agent 10 (γ=4.5) represents a very conservative state, where returns are minimal and growth effectively stops.

Table 1
www.frontiersin.org

Table 1. Multi-agent simulation results (10 agents, 2,000 paths each, 20,000 total).

The simulation results also highlight why market behavior gravitates toward γ2.1. At that point, strategies were not only effective in terms of returns but also survivable under stress. By contrast, the formal optimum at γ3.0 gave slightly better scores on the composite metric, but at the cost of narrower margins. The data thus suggest that human decision-makers prefer robustness over squeezing out the last bit of efficiency.

4.2 Empirical validation of convergence

We next examined how quickly different starting conditions approach the long-term estimate. Each agent began from a different initial risk preference and then simulated 10,000 Monte Carlo paths. Table 2 shows that, regardless of whether the process starts from γ0=0.5, 1.0, 3.0, 5.0, or even 10.0, the trajectories converge toward the same neighborhood at approximately γ=2.1.

Table 2
www.frontiersin.org

Table 2. Convergence to empirical estimate from various starting points.

The numerical patterns in the table suggest that convergence is rapid, although the speed depends on the starting point. The Ornstein-Uhlenbeck formula and probability calculation for convergence are given in Equation 4.2 (Øksendal, 2003).

dγt=θγempiricalγtdt+ξdWt,(4.2)

where θ=0.38 describes the reversion speed, ξ=0.15 is the volatility, and the stationary mean is γempirical=2.1 (the empirical estimate from our sample).

4.3 Behavioral regime boundaries

The behavioral transitions (not phase transitions) occur approximately at:

γ1=1.5±0.1: transition from predominantly bang–bang to smoother control (due to regularization);

γ2=4.5±0.2: transition from active to very conservative (near-zero tokenization).

These values emerge from implementation choices and numerical regularization, not from bifurcations in the Hamiltonian (which remains linear in u).

Note that Figure 3 visualizes only the real estate subset used in our study. The empirical estimate γ=2.1±0.38 is obtained from the full multi-asset dataset (real estate, artworks, and patents) and therefore does not assume that patent trajectories coincide with γ=2.1.

4.4 Comparison with real platforms

To demonstrate the practical advantages of our thermodynamic framework, we compare tokenization trajectories from existing platforms against our model’s output at both the empirical estimate from our sample (γ=2.1) and mathematical optimum under our calibration (γ*3.02).

4.5 Sensitivity analysis

To assess robustness, we perform comprehensive sensitivity analysis on key parameters using Monte Carlo simulations (n = 10,000 paths; see Table 3) (Metropolis and Ulam, 1949; Glasserman, 2003; Boyle et al., 1997). See Supplementary Information for detailed results.

Table 3
www.frontiersin.org

Table 3. Sensitivity of key metrics to ±20% variations.

The analysis underscores that both the empirical estimate from our sample (γ=2.1) and mathematical optimum under our calibration (γ*3.02) represent stable operating points, with minimal sensitivity to moderate parameter variations.

4.6 Implementation algorithm

Implementation involves three components: (i) valuation oracles estimating patent value using gradient boosting and transformers, (ii) HJB solvers using the thermodynamic transformation achieving 47 ms latency on NVIDIA RTX 3090, and (iii) blockchain infrastructure for token management with sub-10s total latency. See Supplementary Information for full algorithm with ten-agent verification.

Monte Carlo simulations (n = 10,000 paths per agent) quantify the framework’s effectiveness (see Table 4) (Metropolis and Ulam, 1949; Glasserman, 2003; Boyle et al., 1997). Risk-neutral strategies yield highest expected returns but face losses below 50% of the initial value in 42% of scenarios. The empirical estimate from our sample (γ=2.1) reduces this probability to 10% while achieving 92% success rate. The mathematical optimum under our specific calibration (γ*3.02) further reduces downside to 5% but at the cost of lower returns. The Sharpe ratio peaks near γ=0.5 due to high returns, but practical success rate peaks at γ=2.1, explaining the empirical estimate in our sample (Markowitz, 1952; Sharpe, 1994; Merton, 1972).

Table 4
www.frontiersin.org

Table 4. Performance comparison across risk sensitivities.

5 Discussion

5.1 What to take away (readability first)

Our results are consistent with a simple rule-of-thumb: robust tokenization (empirical γ2.1) maximizes survival and stabilizes revenues; the benchmark (γ*3.0) optimizes a composite efficiency score under our calibration. That gap is not a flaw of the math but an adaptation to frictions and uncertainty.

5.2 SDG positioning

Risk-aware schedules reduce volatility, tighten downside, and lower effective fees. In practical terms, this widens access to IP monetization and supports SDG-aligned innovation (e.g., resilient infrastructure and climate-related technologies) without relying on fixed, brittle plans. Linearization makes such schedules operational at platform latencies. The framework developed in this study provides a new way to optimize patent tokenization.

By linking economic risk aversion to the inverse temperature in statistical physics, the nonlinear Hamilton–Jacobi–Bellman equation, normally nonlinear and hard to solve, could be linearized through the transformation Φ=eγV. This step made it possible to compute optimal tokenization rates efficiently, even under uncertainty.

The empirical analysis of 45 tokenized assets shows that in our sample, the maximum likelihood estimate is γ=2.1±0.38. In contrast, mathematical analysis of the composite function F(γ)=S(γ)K(γ)B(γ) places the optimum at γ*[2.9,3.2] (baseline γ*3.02) under our specific calibration with parameters from equity-like markets (Campbell et al., 1997; Fama and French, 2002).

This gap amounts to roughly 25.395% in performance terms. Rather than a flaw, it highlights behavioral factors specific to our dataset. The highest success rate of 92% occurs at γ=2.1 in our sample, even though the theoretical maximum under our calibration is located elsewhere. Our multi-agent simulations support this interpretation.

With ten heterogeneous agents and 20,000 simulated paths, the system repeatedly converged to the empirical estimate.

Agent 5, positioned at γ=2.1, achieved strong survival outcomes, with 92% of trajectories meeting the revenue threshold. Agent 7, at γ*3.02, maximized the composite function under our calibration but with slightly different trade-offs.

This pattern is consistent with behavioral finance insights: loss aversion and evolutionary fitness matter as much as formal optimization (Kahneman and Tversky, 1979; Barberis, 2013).

Computational benchmarks show an 850× speedup (from 40 s to 47 ms on NVIDIA RTX 3090), making real-time applications feasible. The broader implication is that mathematical optima under specific calibrations are not always the points observed in empirical samples. Instead, systems under uncertainty tend to evolve toward strategies that balance multiple objectives in practice.

In our case, the estimate at γ=2.1 reflects such balance in our specific dataset.

For future models, explicitly incorporating behavioral anchors and regulatory frictions will be essential if predictions are to match practice. Finally, the approach suggests a wider lesson.

The structures uncovered here are not unique to patents: they mirror principles found in physics, biology, and information theory.

The optimal behavior in complex systems appears to follow universal variational principles, balancing efficiency with survival.

From a practical angle, the numbers are clear: revenue volatility went down substantially, fees dropped to almost negligible levels, and survival rates improved. While these outcomes are encouraging, they are not presented as universal laws. They simply illustrate what can be achieved in our dataset under the chosen calibration. Future work may well find different balances once larger samples or other asset classes are considered. This makes it a candidate for broader application beyond patents to other illiquid assets such as real estate or fine art and positions thermodynamics as a useful lens for understanding financial decision-making under uncertainty.

5.2.1 Data limitations

Our empirical estimate of γ=2.1 derives from 45 tokenization trajectories after cleaning an initial dataset of 57 assets. This sample size, while reasonable for exploratory analysis, raises concerns about statistical power and representativeness. The heterogeneity of assets (37 real estate, 12 artworks, 8 patents) further complicates generalization. Moreover, the sample may not capture the full diversity of tokenization practices, as it is limited to successful cases from specific platforms.

5.2.2 Selection bias

The 45 trajectories represent successful tokenizations, creating survivorship bias. Failed attempts are absent, potentially distorting estimates. This bias could overestimate the effectiveness of moderate strategies, as extreme approaches that failed are not observed. To mitigate this, future studies should include data on unsuccessful tokenizations, though such data are often unavailable.

5.2.3 Model misspecification

The geometric Brownian motion assumption ignores jumps from litigation or regulatory events. Incorporating jumps could alter the optimum. For instance, discrete shocks like patent invalidation occur with non-negligible probability and could shift the optimal γ toward more conservative values. Hybrid models combining continuous diffusion with jump processes are a promising extension, though they would sacrifice some computational efficiency.

5.2.4 Market microstructure

Illiquidity and limited market depth in patent tokens invalidate frictionless assumptions. Current trading volumes are low, leading to price impacts from large token releases. This could make aggressive strategies less viable in practice, explaining part of the deviation from the mathematical optimum. Future work should incorporate liquidity constraints into the optimization framework.

5.2.5 Regulatory uncertainty

Jurisdictional variations create compliance costs that favor conservative strategies. For example, in stringent regimes like the EU, aggressive tokenization might trigger securities regulations, increasing costs. This regulatory friction could explain why the sample estimate is lower than the theoretical optimum. Modeling regulatory scores as part of the objective function is a potential enhancement.

5.3 Behavioral factors

Endowment effects and loss aversion lead to conservative tokenization. Patent holders may overvalue their IP, preferring to retain more control than mathematically optimal. This aligns with prospect theory, where losses loom larger than gains (Kahneman and Tversky, 1979; Barberis, 2013). The deviation in our sample may reflect these biases, suggesting that behavioral adjustments to the model could better match empirical data.

5.3.1 Network effects

Conformity to common standards locks in the observed estimate. Early successful tokenizations were clustered around moderate γ, leading subsequent participants to adopt similar values. This path dependence creates a coordination equilibrium at γ2.1, even if not optimal under our calibration. Breaking this lock-in might require regulatory incentives or platform innovations. Platform dynamics follow established patterns in multi-sided markets (Hagiu and Wright, 2015; Rochet and Tirole, 2006; Evans and Schmalensee, 2008; Rysman, 2009). For extended discussion, see Supplementary Information.

6 Conclusion

Our analysis suggests that risk-sensitive optimization can make patent tokenization both more realistic and more efficient. The link to thermodynamics is not a metaphor but a technical step: with the risk-sensitive sign convention (negative quadratic gradient term) in the Hamilton–Jacobi–Bellman equation (Fleming and Soner, 2006), the exponential transform Φ=eγV cancels the nonlinearity and yields a linear PDE in Φ for any fixed control. The optimal policy then follows from pointwise maximization (threshold/bang–bang, free-boundary structure). In practice, this cuts solve times from seconds to milliseconds—approximately an 850× speedup on NVIDIA RTX 3090—bringing real-time scheduling within reach.

The simulations reveal a consistent pattern as the maximum of the multi-objective function lies at γ*[2.9,3.2] (baseline γ*3.02) under our specific calibration with parameters from equity-like markets (Campbell et al., 1997; Fama and French, 2002), yet the empirical estimate from our sample is γ=2.1±0.38. This difference reflects behavioral factors specific to our dataset where the sample achieves high success rates. In our agent-based analysis, the point at γ=2.1 achieved the highest success rate, while the theoretical optimum under our calibration achieved the best composite score but with different trade-offs.

Monte Carlo experiments underline the practical implications. Transaction costs fall by about 45%, success rates improve from roughly 70%–92%, and downside risk is cut to approximately 10%. The estimate at γ=2.1 thus represents a practical balance in our sample: aggressive enough to generate returns but conservative enough to survive adverse conditions. We count a path as successful if cumulative revenue 0.8 (relative to X0=1). All quantities are normalized to X0=1; time is scaled to [0,1] (t/T).

Looking ahead, the framework offers a quantitative foundation for the next generation of tokenization systems. As blockchain infrastructures scale and regulation stabilizes, the same principles could be extended to other illiquid assets such as real estate or artworks. The broader lesson is that empirical samples, like adaptive systems in biology, may evolve toward solutions that minimize free energy and preserve integrity under uncertainty.

In conclusion, this study demonstrates how thermodynamic reasoning can bridge theory and practice in innovation finance. The same universal variational principles that govern physical and biological systems appear to structure financial decision-making as well. This insight opens new opportunities for research at the intersection of financial engineering, statistical physics, and computational optimization.

Data availability statement

The datasets generated and analyzed in this study contain proprietary information from tokenized assets and are therefore not publicly available; they can be obtained from the corresponding author on reasonable request.

Author contributions

AP: Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Software, Validation, Visualization, Writing – original draft, Writing – review and editing.

Funding

The author declares that no financial support was received for the research and/or publication of this article.

Conflict of interest

The author declares that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author declares that no Generative AI was used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Akcigit, U., and Kerr, W. R. (2018). Growth through heterogeneous innovations. J. Political Econ. 126 (4), 1374–1443. doi:10.1086/697901

CrossRef Full Text | Google Scholar

Arora, A., Ceccagnoli, M., and Cohen, W. M. (2008). R&D and the patent premium. Int. J. Industrial Organ. 26 (5), 1153–1179. doi:10.1016/j.ijindorg.2007.11.004

CrossRef Full Text | Google Scholar

Barberis, N. C. (2013). Thirty years of prospect theory in economics: a review and assessment. J. Econ. Perspect. 27 (1), 173–196. doi:10.1257/jep.27.1.173

CrossRef Full Text | Google Scholar

Baum, A. (2021). Tokenization: the future of real estate investment? Oxford future of real estate initiative. Available online at: https://www.sbs.ox.ac.uk/research/oxford-future-real-estate-initiative.

Google Scholar

Benedetti, H., and Kostovetsky, L. (2021). Digital tulips? Returns to investors in initial coin offerings. J. Corp. Finance 66, 101786. doi:10.1016/j.jcorpfin.2020.101786

CrossRef Full Text | Google Scholar

Bessen, J., and Meurer, M. J. (2008). Patent failure: how judges, bureaucrats, and lawyers put innovators at risk. Princeton University Press. doi:10.1515/9781400828692

CrossRef Full Text | Google Scholar

Bloom, N., Jones, C. I., Van Reenen, J., and Webb, M. (2020). Are ideas getting harder to find? Am. Econ. Rev. 110 (4), 1104–1144. doi:10.1257/aer.20180338

CrossRef Full Text | Google Scholar

Boldrin, M., and Levine, D. K. (2008). Against intellectual monopoly. Cambridge University Press. doi:10.1017/CBO9781139150900

CrossRef Full Text | Google Scholar

Boyle, P., Broadie, M., and Glasserman, P. (1997). Monte carlo methods for security pricing. J. Econ. Dyn. Control 21 (8–9), 1267–1321. doi:10.1016/S0165-1889(97)00028-6

CrossRef Full Text | Google Scholar

Campbell, J. Y., Lo, A. W., and MacKinlay, A. C. (1997). The econometrics of financial markets. Princeton University Press. Available online at: https://press.princeton.edu/books/hardcover/9780691043012/the-econometrics-of-financial-markets.

Google Scholar

Catalini, C., and Gans, J. S. (2018). Initial coin offerings and the value of crypto tokens. NBER Working Paper, No. 24418. doi:10.3386/w24418

CrossRef Full Text | Google Scholar

Chen, Y. (2018). Blockchain tokens and the potential democratization of entrepreneurship and innovation. Bus. Horizons 61 (4), 567–575. doi:10.1016/j.bushor.2018.03.006

CrossRef Full Text | Google Scholar

Cong, L. W., and He, Z. (2019). Blockchain disruption and smart contracts. Rev. Financial Stud. 32 (5), 1754–1797. doi:10.1093/rfs/hhz007

CrossRef Full Text | Google Scholar

Deloitte (2023). Tokenizing intellectual property: current implementations. Deloitte Insights. Available online at: https://www2.deloitte.com/insights/.

Google Scholar

European Patent Office (EPO) (2023). Economic impact of patents and trademarks. Munich: EPO. Available online at: https://www.epo.org/en/news-events/news/2023/20230608.

Google Scholar

Evans, D. S., and Schmalensee, R. (2008). Markets with two-sided platforms. In Issues in Competition Law and Policy. (American Bar Association) 1, 667–693.

Google Scholar

Fama, E. F., and French, K. R. (2002). The equity premium. J. Finance 57 (2), 637–659. doi:10.1111/1540-6261.00437

CrossRef Full Text | Google Scholar

Fleming, W. H., and Soner, H. M. (2006). Controlled markov processes and viscosity solutions. 2nd ed. Springer. doi:10.1007/978-0-387-44612-4

CrossRef Full Text | Google Scholar

Friston, K. (2010). The free-energy principle: a unified brain theory? Nat. Rev. Neurosci. 11 (2), 127–138. doi:10.1038/nrn2787

PubMed Abstract | CrossRef Full Text | Google Scholar

Frontiers (2025). Blockchain and tokenomics for sustainable development, volume II (research topic). Available online at: https://www.frontiersin.org/research-topics/70641/blockchain-and-tokenomics-for-sustainable-development-volume-ii.

Google Scholar

Gambardella, A., Giuri, P., and Luzzi, A. (2007). The market for patents in Europe. Res. Policy 36 (8), 1163–1183. doi:10.1016/j.respol.2007.07.006

CrossRef Full Text | Google Scholar

Glasserman, P. (2003). Monte carlo methods in financial engineering. Springer. doi:10.1007/978-0-387-21617-8

CrossRef Full Text | Google Scholar

Hagiu, A., and Wright, J. (2015). Multi-sided platforms. Int. J. Industrial Organ. 43, 162–174. doi:10.1016/j.ijindorg.2015.03.003

CrossRef Full Text | Google Scholar

Howell, S. T., Niessner, M., and Yermack, D. (2020). Initial coin offerings: financing growth with cryptocurrency token sales. Rev. Financial Stud. 33 (9), 3925–3974. doi:10.1093/rfs/hhz131

CrossRef Full Text | Google Scholar

IBM (2022). Patent tokenization on hyperledger: results and lessons. IBM Blockchain Blog. Available online at: https://www.ibm.com/blogs/blockchain/.

Google Scholar

IPwe (2024). Global patent marketplace metrics. Available online at: https://ipwe.com/.

Google Scholar

Jaynes, E. T. (1957). Information theory and statistical mechanics. Phys. Rev. 106 (4), 620–630. doi:10.1103/PhysRev.106.620

CrossRef Full Text | Google Scholar

Jones, C. I., and Williams, J. C. (2000). Too much of a good thing? The economics of investment in R&D. J. Econ. Growth 5 (1), 65–85. doi:10.1023/A:1009826304308

CrossRef Full Text | Google Scholar

Kahneman, D., and Tversky, A. (1979). Prospect theory: an analysis of decision under risk. Econometrica 47 (2), 263–291. doi:10.2307/1914185

CrossRef Full Text | Google Scholar

Kappen, H. J. (2005). Linear theory for control of nonlinear stochastic systems. Phys. Rev. Lett. 95 (20), 200201. doi:10.1103/PhysRevLett.95.200201

PubMed Abstract | CrossRef Full Text | Google Scholar

Kortum, S., and Lerner, J. (2000). Assessing the contribution of venture capital to innovation. RAND J. Econ. 31 (4), 674–692. doi:10.2307/2696354

CrossRef Full Text | Google Scholar

Kräussl, R., and Whitaker, A. (2020). Blockchain, fractional ownership, and the future of creative work. SSRN Work. Pap. No. 3100389. doi:10.2139/ssrn.3100389

CrossRef Full Text | Google Scholar

Lanjouw, J. O., and Schankerman, M. (2004). Patent quality and research productivity: measuring innovation with multiple indicators. Econ. J. 114 (495), 441–465. doi:10.1111/j.1468-0297.2004.00216.x

CrossRef Full Text | Google Scholar

Lemley, M. A., and Shapiro, C. (2005). Probabilistic patents. J. Econ. Perspect. 19 (2), 75–98. doi:10.1257/0895330054048650

CrossRef Full Text | Google Scholar

Maecenas (2024). Fine art tokenization platform. Available online at: https://www.maecenas.co/.

Google Scholar

Malinova, K., and Park, A. (2017). Market design with blockchain technology. SSRN Work. Pap. No. 2785626. doi:10.2139/ssrn.2785626

CrossRef Full Text | Google Scholar

Markowitz, H. (1952). Portfolio selection. J. Finance 7 (1), 77–91. doi:10.1111/j.1540-6261.1952.tb01525.x

CrossRef Full Text | Google Scholar

Merton, R. C. (1972). An analytic derivation of the efficient portfolio frontier. J. Financial Quantitative Analysis 7 (4), 1851–1872. doi:10.2307/2329621

CrossRef Full Text | Google Scholar

Metropolis, N., and Ulam, S. (1949). The Monte Carlo method. J. Am. Stat. Assoc. 44 (247), 335–341. doi:10.1080/01621459.1949.10483310

PubMed Abstract | CrossRef Full Text | Google Scholar

Øksendal, B. (2003). Stochastic differential equations: an introduction with applications. 6th ed. Springer. doi:10.1007/978-3-642-14394-6

CrossRef Full Text | Google Scholar

RealT (2024). Tokenized real estate on blockchain. Available online at: https://www.realt.co/.

Google Scholar

Rochet, J.-C., and Tirole, J. (2006). Two-sided markets: a progress report. RAND J. Econ. 37 (3), 645–667. doi:10.1111/j.1756-2171.2006.tb00036.x

CrossRef Full Text | Google Scholar

Rysman, M. (2009). The economics of two-sided markets. J. Econ. Perspect. 23 (3), 125–143. doi:10.1257/jep.23.3.125

CrossRef Full Text | Google Scholar

Scotchmer, S. (2004). Innovation and incentives. MIT Press. Available online at: https://mitpress.mit.edu/9780262693431/innovation-and-incentives/.

Google Scholar

Sharpe, W. F. (1994). The sharpe ratio. J. Portfolio Manag. 21 (1), 49–58. doi:10.3905/jpm.1994.409501

CrossRef Full Text | Google Scholar

Todorov, E. (2009). Efficient computation of optimal actions. Proc. Natl. Acad. Sci. U. S. A. 106 (28), 11478–11483. doi:10.1073/pnas.0710743106

PubMed Abstract | CrossRef Full Text | Google Scholar

Touchette, H. (2009). The large deviation approach to statistical mechanics. Phys. Rep. 478 (1–3), 1–69. doi:10.1016/j.physrep.2009.05.002

CrossRef Full Text | Google Scholar

United States Patent and Trademark Office (USPTO) (2024). Patent statistics report 2024. Washington, DC: USPTO. Available online at: https://www.uspto.gov/dashboard/patents/.

Google Scholar

World Intellectual Property Organization (WIPO) (2024). World intellectual property indicators 2024. Geneva: WIPO. doi:10.34667/tind.50133

CrossRef Full Text | Google Scholar

Yermack, D. (2017). Corporate governance and blockchains. Rev. Finance 21 (1), 7–31. doi:10.1093/rof/rfw074

CrossRef Full Text | Google Scholar

Keywords: patent tokenization, risk-sensitive control, Hamilton–Jacobi–Bellman equation, thermodynamics, blockchain economics, sustainable development goals, innovation finance, stochastic optimization

Citation: Peters A (2025) Thermodynamic control of patent tokenization for sustainable development. Front. Blockchain 8:1648418. doi: 10.3389/fbloc.2025.1648418

Received: 17 June 2025; Accepted: 31 October 2025;
Published: 03 December 2025.

Edited by:

Claudio Schifanella, University of Turin, Italy

Reviewed by:

Szymon Łukaszyk, Łukaszyk Patent Attorneys, Poland
Issa Bamia, African Institute for Mathematical Sciences, Cameroon

Copyright © 2025 Peters. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Andreas Peters, YS5wZXRlcnM4MUBpY2xvdWQuY29t

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.