Fixed TeX mistake in expt2

This commit is contained in:
2024-06-18 14:12:26 +01:00
parent cb74559222
commit 8d81ea2008

View File

@@ -52,33 +52,33 @@ This is a testing sandbox for developing various methods of injecting symbolic k
- Multiple attempts carried out: - Multiple attempts carried out:
- Simple penalties. Variations tested include: - Simple penalties. Variations tested include:
```math ```math
Loss = ( Softplus( -m ) + 1 ) * Smooth_L1_Loss Loss = ( Softplus( -m ) + 1 ) * SmoothL1Loss
``` ```
```math ```math
Loss = ( Relu( -m ) + 1 ) * Smooth_L1_Loss Loss = ( Relu( -m ) + 1 ) * SmoothL1Loss
``` ```
```math ```math
Loss = ( \frac{1}{Sech(|r|)} + 1 ) * Smooth_L1_Loss Loss = ( \frac{1}{Sech(|r|)} + 1 ) * SmoothL1Loss
``` ```
```math ```math
Loss = ( {r}^2 + 1) * Smooth_L1_Loss Loss = ( {r}^2 + 1) * SmoothL1Loss
``` ```
- Adaptive, self training penalties tuned by various methods. Best method found was optimisation by a random forest regressor. These tunable variants include: - Adaptive, self training penalties tuned by various methods. Best method found was optimisation by a random forest regressor. These tunable variants include:
```math ```math
Loss = ( Softplus( \alpha * -m ) + 1 ) * Smooth_L1_Loss Loss = ( Softplus( \alpha * -m ) + 1 ) * SmoothL1Loss
``` ```
```math ```math
Loss = ( Relu( \alpha * -m ) + 1 ) * Smooth_L1_Loss Loss = ( Relu( \alpha * -m ) + 1 ) * SmoothL1Loss
``` ```
```math ```math
Loss = ( \frac{ 1 }{ Sech( \alpha * |r| ) } + 1 ) * Smooth_L1_Loss Loss = ( \frac{ 1 }{ Sech( \alpha * |r| ) } + 1 ) * SmoothL1Loss
``` ```
```math ```math
Loss = ( \alpha * { r }^2 + 1) * Smooth_L1_Loss Loss = ( \alpha * { r }^2 + 1) * SmoothL1Loss
``` ```
- Final adaptive semantic loss function tested was the following: - Final adaptive semantic loss function tested was the following:
```math ```math
Loss = ( \alpha * { r }^2 + 1) * ( \frac{ 1 }{ \beta } * log( 1 + exp( \beta * \gamma * -m ) ) + 1 ) * Smooth_L1_Loss Loss = ( \alpha * { r }^2 + 1) * ( \frac{ 1 }{ \beta } * log( 1 + exp( \beta * \gamma * -m ) ) + 1 ) * SmoothL1Loss
``` ```
### Results ### Results