mirror of https://github.com/coder/coder.git
fix(coderd): use insights for DAUs, simplify metricscache (#12775)
Fixes #12134 Fixes https://github.com/coder/customers/issues/384 Refs #12122
This commit is contained in:
parent
5d82a78d4c
commit
421bf7e785
|
@ -366,8 +366,8 @@ func New(options *Options) *API {
|
||||||
options.Database,
|
options.Database,
|
||||||
options.Logger.Named("metrics_cache"),
|
options.Logger.Named("metrics_cache"),
|
||||||
metricscache.Intervals{
|
metricscache.Intervals{
|
||||||
TemplateDAUs: options.MetricsCacheRefreshInterval,
|
TemplateBuildTimes: options.MetricsCacheRefreshInterval,
|
||||||
DeploymentStats: options.AgentStatsRefreshInterval,
|
DeploymentStats: options.AgentStatsRefreshInterval,
|
||||||
},
|
},
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
|
@ -32,14 +32,19 @@ const insightsTimeLayout = time.RFC3339
|
||||||
// @Success 200 {object} codersdk.DAUsResponse
|
// @Success 200 {object} codersdk.DAUsResponse
|
||||||
// @Router /insights/daus [get]
|
// @Router /insights/daus [get]
|
||||||
func (api *API) deploymentDAUs(rw http.ResponseWriter, r *http.Request) {
|
func (api *API) deploymentDAUs(rw http.ResponseWriter, r *http.Request) {
|
||||||
ctx := r.Context()
|
|
||||||
if !api.Authorize(r, rbac.ActionRead, rbac.ResourceDeploymentValues) {
|
if !api.Authorize(r, rbac.ActionRead, rbac.ResourceDeploymentValues) {
|
||||||
httpapi.Forbidden(rw)
|
httpapi.Forbidden(rw)
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
vals := r.URL.Query()
|
api.returnDAUsInternal(rw, r, nil)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (api *API) returnDAUsInternal(rw http.ResponseWriter, r *http.Request, templateIDs []uuid.UUID) {
|
||||||
|
ctx := r.Context()
|
||||||
|
|
||||||
p := httpapi.NewQueryParamParser()
|
p := httpapi.NewQueryParamParser()
|
||||||
|
vals := r.URL.Query()
|
||||||
tzOffset := p.Int(vals, 0, "tz_offset")
|
tzOffset := p.Int(vals, 0, "tz_offset")
|
||||||
p.ErrorExcessParams(vals)
|
p.ErrorExcessParams(vals)
|
||||||
if len(p.Errors) > 0 {
|
if len(p.Errors) > 0 {
|
||||||
|
@ -50,12 +55,41 @@ func (api *API) deploymentDAUs(rw http.ResponseWriter, r *http.Request) {
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
_, resp, _ := api.metricsCache.DeploymentDAUs(tzOffset)
|
loc := time.FixedZone("", tzOffset*3600)
|
||||||
if resp == nil || resp.Entries == nil {
|
// If the time is 14:01 or 14:31, we still want to include all the
|
||||||
httpapi.Write(ctx, rw, http.StatusOK, &codersdk.DAUsResponse{
|
// data between 14:00 and 15:00. Our rollups buckets are 30 minutes
|
||||||
Entries: []codersdk.DAUEntry{},
|
// so this works nicely. It works just as well for 23:59 as well.
|
||||||
|
nextHourInLoc := time.Now().In(loc).Truncate(time.Hour).Add(time.Hour)
|
||||||
|
// Always return 60 days of data (2 months).
|
||||||
|
sixtyDaysAgo := nextHourInLoc.In(loc).Truncate(24*time.Hour).AddDate(0, 0, -60)
|
||||||
|
|
||||||
|
rows, err := api.Database.GetTemplateInsightsByInterval(ctx, database.GetTemplateInsightsByIntervalParams{
|
||||||
|
StartTime: sixtyDaysAgo,
|
||||||
|
EndTime: nextHourInLoc,
|
||||||
|
IntervalDays: 1,
|
||||||
|
TemplateIDs: templateIDs,
|
||||||
|
})
|
||||||
|
if err != nil {
|
||||||
|
if httpapi.Is404Error(err) {
|
||||||
|
httpapi.ResourceNotFound(rw)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
httpapi.Write(ctx, rw, http.StatusInternalServerError, codersdk.Response{
|
||||||
|
Message: "Internal error fetching DAUs.",
|
||||||
|
Detail: err.Error(),
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
resp := codersdk.DAUsResponse{
|
||||||
|
TZHourOffset: tzOffset,
|
||||||
|
Entries: make([]codersdk.DAUEntry, 0, len(rows)),
|
||||||
|
}
|
||||||
|
for _, row := range rows {
|
||||||
|
resp.Entries = append(resp.Entries, codersdk.DAUEntry{
|
||||||
|
Date: row.StartTime.Format(time.DateOnly),
|
||||||
|
Amount: int(row.ActiveUsers),
|
||||||
})
|
})
|
||||||
return
|
|
||||||
}
|
}
|
||||||
httpapi.Write(ctx, rw, http.StatusOK, resp)
|
httpapi.Write(ctx, rw, http.StatusOK, resp)
|
||||||
}
|
}
|
||||||
|
|
|
@ -39,25 +39,25 @@ import (
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestDeploymentInsights(t *testing.T) {
|
func TestDeploymentInsights(t *testing.T) {
|
||||||
t.Skipf("This test is flaky: https://github.com/coder/coder/issues/12509")
|
|
||||||
|
|
||||||
t.Parallel()
|
t.Parallel()
|
||||||
|
|
||||||
clientTz, err := time.LoadLocation("America/Chicago")
|
clientTz, err := time.LoadLocation("America/Chicago")
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
|
|
||||||
db, ps := dbtestutil.NewDB(t)
|
db, ps := dbtestutil.NewDB(t, dbtestutil.WithDumpOnFailure())
|
||||||
logger := slogtest.Make(t, nil)
|
logger := slogtest.Make(t, nil)
|
||||||
|
rollupEvents := make(chan dbrollup.Event)
|
||||||
client := coderdtest.New(t, &coderdtest.Options{
|
client := coderdtest.New(t, &coderdtest.Options{
|
||||||
Database: db,
|
Database: db,
|
||||||
Pubsub: ps,
|
Pubsub: ps,
|
||||||
Logger: &logger,
|
Logger: &logger,
|
||||||
IncludeProvisionerDaemon: true,
|
IncludeProvisionerDaemon: true,
|
||||||
AgentStatsRefreshInterval: time.Millisecond * 50,
|
AgentStatsRefreshInterval: time.Millisecond * 100,
|
||||||
DatabaseRolluper: dbrollup.New(
|
DatabaseRolluper: dbrollup.New(
|
||||||
logger.Named("dbrollup"),
|
logger.Named("dbrollup").Leveled(slog.LevelDebug),
|
||||||
db,
|
db,
|
||||||
dbrollup.WithInterval(time.Millisecond*100),
|
dbrollup.WithInterval(time.Millisecond*100),
|
||||||
|
dbrollup.WithEventChannel(rollupEvents),
|
||||||
),
|
),
|
||||||
})
|
})
|
||||||
|
|
||||||
|
@ -75,57 +75,51 @@ func TestDeploymentInsights(t *testing.T) {
|
||||||
workspace := coderdtest.CreateWorkspace(t, client, user.OrganizationID, template.ID)
|
workspace := coderdtest.CreateWorkspace(t, client, user.OrganizationID, template.ID)
|
||||||
coderdtest.AwaitWorkspaceBuildJobCompleted(t, client, workspace.LatestBuild.ID)
|
coderdtest.AwaitWorkspaceBuildJobCompleted(t, client, workspace.LatestBuild.ID)
|
||||||
|
|
||||||
|
ctx := testutil.Context(t, testutil.WaitLong)
|
||||||
|
|
||||||
|
// Pre-check, no permission issues.
|
||||||
|
daus, err := client.DeploymentDAUs(ctx, codersdk.TimezoneOffsetHour(clientTz))
|
||||||
|
require.NoError(t, err)
|
||||||
|
|
||||||
_ = agenttest.New(t, client.URL, authToken)
|
_ = agenttest.New(t, client.URL, authToken)
|
||||||
resources := coderdtest.AwaitWorkspaceAgents(t, client, workspace.ID)
|
resources := coderdtest.NewWorkspaceAgentWaiter(t, client, workspace.ID).Wait()
|
||||||
ctx, cancel := context.WithTimeout(context.Background(), testutil.WaitLong)
|
|
||||||
defer cancel()
|
|
||||||
|
|
||||||
daus, err := client.DeploymentDAUs(context.Background(), codersdk.TimezoneOffsetHour(clientTz))
|
|
||||||
require.NoError(t, err)
|
|
||||||
|
|
||||||
res, err := client.Workspaces(ctx, codersdk.WorkspaceFilter{})
|
|
||||||
require.NoError(t, err)
|
|
||||||
assert.NotZero(t, res.Workspaces[0].LastUsedAt)
|
|
||||||
|
|
||||||
conn, err := workspacesdk.New(client).
|
conn, err := workspacesdk.New(client).
|
||||||
DialAgent(ctx, resources[0].Agents[0].ID, &workspacesdk.DialAgentOptions{
|
DialAgent(ctx, resources[0].Agents[0].ID, &workspacesdk.DialAgentOptions{
|
||||||
Logger: slogtest.Make(t, nil).Named("tailnet"),
|
Logger: slogtest.Make(t, nil).Named("dialagent"),
|
||||||
})
|
})
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
defer func() {
|
defer conn.Close()
|
||||||
_ = conn.Close()
|
|
||||||
}()
|
|
||||||
|
|
||||||
sshConn, err := conn.SSHClient(ctx)
|
sshConn, err := conn.SSHClient(ctx)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
_ = sshConn.Close()
|
defer sshConn.Close()
|
||||||
|
|
||||||
|
sess, err := sshConn.NewSession()
|
||||||
|
require.NoError(t, err)
|
||||||
|
defer sess.Close()
|
||||||
|
|
||||||
|
r, w := io.Pipe()
|
||||||
|
defer r.Close()
|
||||||
|
defer w.Close()
|
||||||
|
sess.Stdin = r
|
||||||
|
sess.Stdout = io.Discard
|
||||||
|
err = sess.Start("cat")
|
||||||
|
require.NoError(t, err)
|
||||||
|
|
||||||
|
for {
|
||||||
|
select {
|
||||||
|
case <-ctx.Done():
|
||||||
|
require.Fail(t, "timed out waiting for deployment daus to update", daus)
|
||||||
|
case <-rollupEvents:
|
||||||
|
}
|
||||||
|
|
||||||
wantDAUs := &codersdk.DAUsResponse{
|
|
||||||
TZHourOffset: codersdk.TimezoneOffsetHour(clientTz),
|
|
||||||
Entries: []codersdk.DAUEntry{
|
|
||||||
{
|
|
||||||
Date: time.Now().In(clientTz).Format("2006-01-02"),
|
|
||||||
Amount: 1,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}
|
|
||||||
require.Eventuallyf(t, func() bool {
|
|
||||||
daus, err = client.DeploymentDAUs(ctx, codersdk.TimezoneOffsetHour(clientTz))
|
daus, err = client.DeploymentDAUs(ctx, codersdk.TimezoneOffsetHour(clientTz))
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
return len(daus.Entries) > 0
|
if len(daus.Entries) > 0 && daus.Entries[len(daus.Entries)-1].Amount > 0 {
|
||||||
},
|
break
|
||||||
testutil.WaitShort, testutil.IntervalFast,
|
}
|
||||||
"deployment daus never loaded",
|
}
|
||||||
)
|
|
||||||
gotDAUs, err := client.DeploymentDAUs(ctx, codersdk.TimezoneOffsetHour(clientTz))
|
|
||||||
require.NoError(t, err)
|
|
||||||
require.Equal(t, gotDAUs, wantDAUs)
|
|
||||||
|
|
||||||
template, err = client.Template(ctx, template.ID)
|
|
||||||
require.NoError(t, err)
|
|
||||||
|
|
||||||
res, err = client.Workspaces(ctx, codersdk.WorkspaceFilter{})
|
|
||||||
require.NoError(t, err)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestUserActivityInsights_SanityCheck(t *testing.T) {
|
func TestUserActivityInsights_SanityCheck(t *testing.T) {
|
||||||
|
|
|
@ -1,93 +0,0 @@
|
||||||
package metricscache
|
|
||||||
|
|
||||||
import (
|
|
||||||
"testing"
|
|
||||||
|
|
||||||
"github.com/stretchr/testify/require"
|
|
||||||
)
|
|
||||||
|
|
||||||
func TestClosest(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
testCases := []struct {
|
|
||||||
Name string
|
|
||||||
Keys []int
|
|
||||||
Input int
|
|
||||||
Expected int
|
|
||||||
NotFound bool
|
|
||||||
}{
|
|
||||||
{
|
|
||||||
Name: "Empty",
|
|
||||||
Input: 10,
|
|
||||||
NotFound: true,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Name: "Equal",
|
|
||||||
Keys: []int{1, 2, 3, 4, 5, 6, 10, 12, 15},
|
|
||||||
Input: 10,
|
|
||||||
Expected: 10,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Name: "ZeroOnly",
|
|
||||||
Keys: []int{0},
|
|
||||||
Input: 10,
|
|
||||||
Expected: 0,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Name: "NegativeOnly",
|
|
||||||
Keys: []int{-10, -5},
|
|
||||||
Input: 10,
|
|
||||||
Expected: -5,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Name: "CloseBothSides",
|
|
||||||
Keys: []int{-10, -5, 0, 5, 8, 12},
|
|
||||||
Input: 10,
|
|
||||||
Expected: 8,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Name: "CloseNoZero",
|
|
||||||
Keys: []int{-10, -5, 5, 8, 12},
|
|
||||||
Input: 0,
|
|
||||||
Expected: -5,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Name: "CloseLeft",
|
|
||||||
Keys: []int{-10, -5, 0, 5, 8, 12},
|
|
||||||
Input: 20,
|
|
||||||
Expected: 12,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Name: "CloseRight",
|
|
||||||
Keys: []int{-10, -5, 0, 5, 8, 12},
|
|
||||||
Input: -20,
|
|
||||||
Expected: -10,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Name: "ChooseZero",
|
|
||||||
Keys: []int{-10, -5, 0, 5, 8, 12},
|
|
||||||
Input: 2,
|
|
||||||
Expected: 0,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, tc := range testCases {
|
|
||||||
tc := tc
|
|
||||||
t.Run(tc.Name, func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
m := make(map[int]int)
|
|
||||||
for _, k := range tc.Keys {
|
|
||||||
m[k] = k
|
|
||||||
}
|
|
||||||
|
|
||||||
found, _, ok := closest(m, tc.Input)
|
|
||||||
if tc.NotFound {
|
|
||||||
require.False(t, ok, "should not be found")
|
|
||||||
} else {
|
|
||||||
require.True(t, ok)
|
|
||||||
require.Equal(t, tc.Expected, found, "closest")
|
|
||||||
}
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
|
@ -3,15 +3,11 @@ package metricscache
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"database/sql"
|
"database/sql"
|
||||||
"fmt"
|
|
||||||
"math"
|
|
||||||
"sync"
|
"sync"
|
||||||
"sync/atomic"
|
"sync/atomic"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/google/uuid"
|
"github.com/google/uuid"
|
||||||
"golang.org/x/exp/maps"
|
|
||||||
"golang.org/x/exp/slices"
|
|
||||||
"golang.org/x/xerrors"
|
"golang.org/x/xerrors"
|
||||||
|
|
||||||
"cdr.dev/slog"
|
"cdr.dev/slog"
|
||||||
|
@ -22,33 +18,6 @@ import (
|
||||||
"github.com/coder/retry"
|
"github.com/coder/retry"
|
||||||
)
|
)
|
||||||
|
|
||||||
func OnlyDate(t time.Time) string {
|
|
||||||
return t.Format("2006-01-02")
|
|
||||||
}
|
|
||||||
|
|
||||||
// deploymentTimezoneOffsets are the timezones that are cached and supported.
|
|
||||||
// Any non-listed timezone offsets will need to use the closest supported one.
|
|
||||||
var deploymentTimezoneOffsets = []int{
|
|
||||||
0, // UTC - is listed first intentionally.
|
|
||||||
// Shortened list of 4 timezones that should encompass *most* users. Caching
|
|
||||||
// all 25 timezones can be too computationally expensive for large
|
|
||||||
// deployments. This is a stop-gap until more robust fixes can be made for
|
|
||||||
// the deployment DAUs query.
|
|
||||||
-6, 3, 6, 10,
|
|
||||||
|
|
||||||
// -12, -11, -10, -9, -8, -7, -6, -5, -4, -3, -2, -1,
|
|
||||||
// 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12,
|
|
||||||
}
|
|
||||||
|
|
||||||
// templateTimezoneOffsets are the timezones each template will use for it's DAU
|
|
||||||
// calculations. This is expensive as each template needs to do each timezone, so keep this list
|
|
||||||
// very small.
|
|
||||||
var templateTimezoneOffsets = []int{
|
|
||||||
// Only do one for now. If people request more accurate template DAU, we can
|
|
||||||
// fix this. But it adds too much cost, so optimization is needed first.
|
|
||||||
0, // UTC - is listed first intentionally.
|
|
||||||
}
|
|
||||||
|
|
||||||
// Cache holds the template metrics.
|
// Cache holds the template metrics.
|
||||||
// The aggregation queries responsible for these values can take up to a minute
|
// The aggregation queries responsible for these values can take up to a minute
|
||||||
// on large deployments. Even in small deployments, aggregation queries can
|
// on large deployments. Even in small deployments, aggregation queries can
|
||||||
|
@ -59,9 +28,6 @@ type Cache struct {
|
||||||
log slog.Logger
|
log slog.Logger
|
||||||
intervals Intervals
|
intervals Intervals
|
||||||
|
|
||||||
deploymentDAUResponses atomic.Pointer[map[int]codersdk.DAUsResponse]
|
|
||||||
templateDAUResponses atomic.Pointer[map[int]map[uuid.UUID]codersdk.DAUsResponse]
|
|
||||||
templateUniqueUsers atomic.Pointer[map[uuid.UUID]int]
|
|
||||||
templateWorkspaceOwners atomic.Pointer[map[uuid.UUID]int]
|
templateWorkspaceOwners atomic.Pointer[map[uuid.UUID]int]
|
||||||
templateAverageBuildTime atomic.Pointer[map[uuid.UUID]database.GetTemplateAverageBuildTimeRow]
|
templateAverageBuildTime atomic.Pointer[map[uuid.UUID]database.GetTemplateAverageBuildTimeRow]
|
||||||
deploymentStatsResponse atomic.Pointer[codersdk.DeploymentStats]
|
deploymentStatsResponse atomic.Pointer[codersdk.DeploymentStats]
|
||||||
|
@ -71,13 +37,13 @@ type Cache struct {
|
||||||
}
|
}
|
||||||
|
|
||||||
type Intervals struct {
|
type Intervals struct {
|
||||||
TemplateDAUs time.Duration
|
TemplateBuildTimes time.Duration
|
||||||
DeploymentStats time.Duration
|
DeploymentStats time.Duration
|
||||||
}
|
}
|
||||||
|
|
||||||
func New(db database.Store, log slog.Logger, intervals Intervals) *Cache {
|
func New(db database.Store, log slog.Logger, intervals Intervals) *Cache {
|
||||||
if intervals.TemplateDAUs <= 0 {
|
if intervals.TemplateBuildTimes <= 0 {
|
||||||
intervals.TemplateDAUs = time.Hour
|
intervals.TemplateBuildTimes = time.Hour
|
||||||
}
|
}
|
||||||
if intervals.DeploymentStats <= 0 {
|
if intervals.DeploymentStats <= 0 {
|
||||||
intervals.DeploymentStats = time.Minute
|
intervals.DeploymentStats = time.Minute
|
||||||
|
@ -97,7 +63,7 @@ func New(db database.Store, log slog.Logger, intervals Intervals) *Cache {
|
||||||
wg.Add(1)
|
wg.Add(1)
|
||||||
go func() {
|
go func() {
|
||||||
defer wg.Done()
|
defer wg.Done()
|
||||||
c.run(ctx, "template daus", intervals.TemplateDAUs, c.refreshTemplateDAUs)
|
c.run(ctx, "template build times", intervals.TemplateBuildTimes, c.refreshTemplateBuildTimes)
|
||||||
}()
|
}()
|
||||||
wg.Add(1)
|
wg.Add(1)
|
||||||
go func() {
|
go func() {
|
||||||
|
@ -109,104 +75,7 @@ func New(db database.Store, log slog.Logger, intervals Intervals) *Cache {
|
||||||
return c
|
return c
|
||||||
}
|
}
|
||||||
|
|
||||||
func fillEmptyDays(sortedDates []time.Time) []time.Time {
|
func (c *Cache) refreshTemplateBuildTimes(ctx context.Context) error {
|
||||||
var newDates []time.Time
|
|
||||||
|
|
||||||
for i, ti := range sortedDates {
|
|
||||||
if i == 0 {
|
|
||||||
newDates = append(newDates, ti)
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
|
|
||||||
last := sortedDates[i-1]
|
|
||||||
|
|
||||||
const day = time.Hour * 24
|
|
||||||
diff := ti.Sub(last)
|
|
||||||
for diff > day {
|
|
||||||
if diff <= day {
|
|
||||||
break
|
|
||||||
}
|
|
||||||
last = last.Add(day)
|
|
||||||
newDates = append(newDates, last)
|
|
||||||
diff -= day
|
|
||||||
}
|
|
||||||
|
|
||||||
newDates = append(newDates, ti)
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
|
|
||||||
return newDates
|
|
||||||
}
|
|
||||||
|
|
||||||
type dauRow interface {
|
|
||||||
database.GetTemplateDAUsRow |
|
|
||||||
database.GetDeploymentDAUsRow
|
|
||||||
}
|
|
||||||
|
|
||||||
func convertDAUResponse[T dauRow](rows []T, tzOffset int) codersdk.DAUsResponse {
|
|
||||||
respMap := make(map[time.Time][]uuid.UUID)
|
|
||||||
for _, row := range rows {
|
|
||||||
switch row := any(row).(type) {
|
|
||||||
case database.GetDeploymentDAUsRow:
|
|
||||||
respMap[row.Date] = append(respMap[row.Date], row.UserID)
|
|
||||||
case database.GetTemplateDAUsRow:
|
|
||||||
respMap[row.Date] = append(respMap[row.Date], row.UserID)
|
|
||||||
default:
|
|
||||||
// This should never happen.
|
|
||||||
panic(fmt.Sprintf("%T not acceptable, developer error", row))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
dates := maps.Keys(respMap)
|
|
||||||
slices.SortFunc(dates, func(a, b time.Time) int {
|
|
||||||
if a.Before(b) {
|
|
||||||
return -1
|
|
||||||
} else if a.Equal(b) {
|
|
||||||
return 0
|
|
||||||
}
|
|
||||||
return 1
|
|
||||||
})
|
|
||||||
|
|
||||||
var resp codersdk.DAUsResponse
|
|
||||||
for _, date := range fillEmptyDays(dates) {
|
|
||||||
resp.Entries = append(resp.Entries, codersdk.DAUEntry{
|
|
||||||
// This date is truncated to 00:00:00 of the given day, so only
|
|
||||||
// return date information.
|
|
||||||
Date: OnlyDate(date),
|
|
||||||
Amount: len(respMap[date]),
|
|
||||||
})
|
|
||||||
}
|
|
||||||
resp.TZHourOffset = tzOffset
|
|
||||||
|
|
||||||
return resp
|
|
||||||
}
|
|
||||||
|
|
||||||
func countUniqueUsers(rows []database.GetTemplateDAUsRow) int {
|
|
||||||
seen := make(map[uuid.UUID]struct{}, len(rows))
|
|
||||||
for _, row := range rows {
|
|
||||||
seen[row.UserID] = struct{}{}
|
|
||||||
}
|
|
||||||
return len(seen)
|
|
||||||
}
|
|
||||||
|
|
||||||
func (c *Cache) refreshDeploymentDAUs(ctx context.Context) error {
|
|
||||||
//nolint:gocritic // This is a system service.
|
|
||||||
ctx = dbauthz.AsSystemRestricted(ctx)
|
|
||||||
|
|
||||||
deploymentDAUs := make(map[int]codersdk.DAUsResponse)
|
|
||||||
for _, tzOffset := range deploymentTimezoneOffsets {
|
|
||||||
rows, err := c.database.GetDeploymentDAUs(ctx, int32(tzOffset))
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
deploymentDAUs[tzOffset] = convertDAUResponse(rows, tzOffset)
|
|
||||||
}
|
|
||||||
|
|
||||||
c.deploymentDAUResponses.Store(&deploymentDAUs)
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (c *Cache) refreshTemplateDAUs(ctx context.Context) error {
|
|
||||||
//nolint:gocritic // This is a system service.
|
//nolint:gocritic // This is a system service.
|
||||||
ctx = dbauthz.AsSystemRestricted(ctx)
|
ctx = dbauthz.AsSystemRestricted(ctx)
|
||||||
|
|
||||||
|
@ -216,38 +85,13 @@ func (c *Cache) refreshTemplateDAUs(ctx context.Context) error {
|
||||||
}
|
}
|
||||||
|
|
||||||
var (
|
var (
|
||||||
templateDAUs = make(map[int]map[uuid.UUID]codersdk.DAUsResponse, len(templates))
|
|
||||||
templateUniqueUsers = make(map[uuid.UUID]int)
|
|
||||||
templateWorkspaceOwners = make(map[uuid.UUID]int)
|
templateWorkspaceOwners = make(map[uuid.UUID]int)
|
||||||
templateAverageBuildTimes = make(map[uuid.UUID]database.GetTemplateAverageBuildTimeRow)
|
templateAverageBuildTimes = make(map[uuid.UUID]database.GetTemplateAverageBuildTimeRow)
|
||||||
)
|
)
|
||||||
|
|
||||||
err = c.refreshDeploymentDAUs(ctx)
|
|
||||||
if err != nil {
|
|
||||||
return xerrors.Errorf("deployment daus: %w", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
ids := make([]uuid.UUID, 0, len(templates))
|
ids := make([]uuid.UUID, 0, len(templates))
|
||||||
for _, template := range templates {
|
for _, template := range templates {
|
||||||
ids = append(ids, template.ID)
|
ids = append(ids, template.ID)
|
||||||
for _, tzOffset := range templateTimezoneOffsets {
|
|
||||||
rows, err := c.database.GetTemplateDAUs(ctx, database.GetTemplateDAUsParams{
|
|
||||||
TemplateID: template.ID,
|
|
||||||
TzOffset: int32(tzOffset),
|
|
||||||
})
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
if templateDAUs[tzOffset] == nil {
|
|
||||||
templateDAUs[tzOffset] = make(map[uuid.UUID]codersdk.DAUsResponse)
|
|
||||||
}
|
|
||||||
templateDAUs[tzOffset][template.ID] = convertDAUResponse(rows, tzOffset)
|
|
||||||
if _, set := templateUniqueUsers[template.ID]; !set {
|
|
||||||
// If the uniqueUsers has not been counted yet, set the unique count with the rows we have.
|
|
||||||
// We only need to calculate this once.
|
|
||||||
templateUniqueUsers[template.ID] = countUniqueUsers(rows)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
templateAvgBuildTime, err := c.database.GetTemplateAverageBuildTime(ctx, database.GetTemplateAverageBuildTimeParams{
|
templateAvgBuildTime, err := c.database.GetTemplateAverageBuildTime(ctx, database.GetTemplateAverageBuildTimeParams{
|
||||||
TemplateID: uuid.NullUUID{
|
TemplateID: uuid.NullUUID{
|
||||||
|
@ -275,8 +119,6 @@ func (c *Cache) refreshTemplateDAUs(ctx context.Context) error {
|
||||||
}
|
}
|
||||||
|
|
||||||
c.templateWorkspaceOwners.Store(&templateWorkspaceOwners)
|
c.templateWorkspaceOwners.Store(&templateWorkspaceOwners)
|
||||||
c.templateDAUResponses.Store(&templateDAUs)
|
|
||||||
c.templateUniqueUsers.Store(&templateUniqueUsers)
|
|
||||||
c.templateAverageBuildTime.Store(&templateAverageBuildTimes)
|
c.templateAverageBuildTime.Store(&templateAverageBuildTimes)
|
||||||
|
|
||||||
return nil
|
return nil
|
||||||
|
@ -359,99 +201,6 @@ func (c *Cache) Close() error {
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *Cache) DeploymentDAUs(offset int) (int, *codersdk.DAUsResponse, bool) {
|
|
||||||
m := c.deploymentDAUResponses.Load()
|
|
||||||
if m == nil {
|
|
||||||
return 0, nil, false
|
|
||||||
}
|
|
||||||
closestOffset, resp, ok := closest(*m, offset)
|
|
||||||
if !ok {
|
|
||||||
return 0, nil, false
|
|
||||||
}
|
|
||||||
return closestOffset, &resp, ok
|
|
||||||
}
|
|
||||||
|
|
||||||
// TemplateDAUs returns an empty response if the template doesn't have users
|
|
||||||
// or is loading for the first time.
|
|
||||||
// The cache will select the closest DAUs response to given timezone offset.
|
|
||||||
func (c *Cache) TemplateDAUs(id uuid.UUID, offset int) (int, *codersdk.DAUsResponse, bool) {
|
|
||||||
m := c.templateDAUResponses.Load()
|
|
||||||
if m == nil {
|
|
||||||
// Data loading.
|
|
||||||
return 0, nil, false
|
|
||||||
}
|
|
||||||
|
|
||||||
closestOffset, resp, ok := closest(*m, offset)
|
|
||||||
if !ok {
|
|
||||||
// Probably no data.
|
|
||||||
return 0, nil, false
|
|
||||||
}
|
|
||||||
|
|
||||||
tpl, ok := resp[id]
|
|
||||||
if !ok {
|
|
||||||
// Probably no data.
|
|
||||||
return 0, nil, false
|
|
||||||
}
|
|
||||||
|
|
||||||
return closestOffset, &tpl, true
|
|
||||||
}
|
|
||||||
|
|
||||||
// closest returns the value in the values map that has a key with the value most
|
|
||||||
// close to the requested key. This is so if a user requests a timezone offset that
|
|
||||||
// we do not have, we return the closest one we do have to the user.
|
|
||||||
func closest[V any](values map[int]V, offset int) (int, V, bool) {
|
|
||||||
if len(values) == 0 {
|
|
||||||
var v V
|
|
||||||
return -1, v, false
|
|
||||||
}
|
|
||||||
|
|
||||||
v, ok := values[offset]
|
|
||||||
if ok {
|
|
||||||
// We have the exact offset, that was easy!
|
|
||||||
return offset, v, true
|
|
||||||
}
|
|
||||||
|
|
||||||
var closest int
|
|
||||||
var closestV V
|
|
||||||
diff := math.MaxInt
|
|
||||||
for k, v := range values {
|
|
||||||
newDiff := abs(k - offset)
|
|
||||||
// Take the closest value that is also the smallest value. We do this
|
|
||||||
// to make the output deterministic
|
|
||||||
if newDiff < diff || (newDiff == diff && k < closest) {
|
|
||||||
// new closest
|
|
||||||
closest = k
|
|
||||||
closestV = v
|
|
||||||
diff = newDiff
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return closest, closestV, true
|
|
||||||
}
|
|
||||||
|
|
||||||
func abs(a int) int {
|
|
||||||
if a < 0 {
|
|
||||||
return -1 * a
|
|
||||||
}
|
|
||||||
return a
|
|
||||||
}
|
|
||||||
|
|
||||||
// TemplateUniqueUsers returns the number of unique Template users
|
|
||||||
// from all Cache data.
|
|
||||||
func (c *Cache) TemplateUniqueUsers(id uuid.UUID) (int, bool) {
|
|
||||||
m := c.templateUniqueUsers.Load()
|
|
||||||
if m == nil {
|
|
||||||
// Data loading.
|
|
||||||
return -1, false
|
|
||||||
}
|
|
||||||
|
|
||||||
resp, ok := (*m)[id]
|
|
||||||
if !ok {
|
|
||||||
// Probably no data.
|
|
||||||
return -1, false
|
|
||||||
}
|
|
||||||
return resp, true
|
|
||||||
}
|
|
||||||
|
|
||||||
func (c *Cache) TemplateBuildTimeStats(id uuid.UUID) codersdk.TemplateBuildTimeStats {
|
func (c *Cache) TemplateBuildTimeStats(id uuid.UUID) codersdk.TemplateBuildTimeStats {
|
||||||
unknown := codersdk.TemplateBuildTimeStats{
|
unknown := codersdk.TemplateBuildTimeStats{
|
||||||
codersdk.WorkspaceTransitionStart: {},
|
codersdk.WorkspaceTransitionStart: {},
|
||||||
|
|
|
@ -19,241 +19,10 @@ import (
|
||||||
"github.com/coder/coder/v2/testutil"
|
"github.com/coder/coder/v2/testutil"
|
||||||
)
|
)
|
||||||
|
|
||||||
func dateH(year, month, day, hour int) time.Time {
|
|
||||||
return time.Date(year, time.Month(month), day, hour, 0, 0, 0, time.UTC)
|
|
||||||
}
|
|
||||||
|
|
||||||
func date(year, month, day int) time.Time {
|
func date(year, month, day int) time.Time {
|
||||||
return time.Date(year, time.Month(month), day, 0, 0, 0, 0, time.UTC)
|
return time.Date(year, time.Month(month), day, 0, 0, 0, 0, time.UTC)
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestCache_TemplateUsers(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
statRow := func(user uuid.UUID, date time.Time) database.InsertWorkspaceAgentStatParams {
|
|
||||||
return database.InsertWorkspaceAgentStatParams{
|
|
||||||
CreatedAt: date,
|
|
||||||
UserID: user,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
var (
|
|
||||||
zebra = uuid.UUID{1}
|
|
||||||
tiger = uuid.UUID{2}
|
|
||||||
)
|
|
||||||
|
|
||||||
type args struct {
|
|
||||||
rows []database.InsertWorkspaceAgentStatParams
|
|
||||||
}
|
|
||||||
type want struct {
|
|
||||||
entries []codersdk.DAUEntry
|
|
||||||
uniqueUsers int
|
|
||||||
}
|
|
||||||
tests := []struct {
|
|
||||||
name string
|
|
||||||
args args
|
|
||||||
tplWant want
|
|
||||||
// dauWant is optional
|
|
||||||
dauWant []codersdk.DAUEntry
|
|
||||||
tzOffset int
|
|
||||||
}{
|
|
||||||
{name: "empty", args: args{}, tplWant: want{nil, 0}},
|
|
||||||
{
|
|
||||||
name: "one hole",
|
|
||||||
args: args{
|
|
||||||
rows: []database.InsertWorkspaceAgentStatParams{
|
|
||||||
statRow(zebra, dateH(2022, 8, 27, 0)),
|
|
||||||
statRow(zebra, dateH(2022, 8, 30, 0)),
|
|
||||||
},
|
|
||||||
},
|
|
||||||
tplWant: want{[]codersdk.DAUEntry{
|
|
||||||
{
|
|
||||||
Date: metricscache.OnlyDate(date(2022, 8, 27)),
|
|
||||||
Amount: 1,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Date: metricscache.OnlyDate(date(2022, 8, 28)),
|
|
||||||
Amount: 0,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Date: metricscache.OnlyDate(date(2022, 8, 29)),
|
|
||||||
Amount: 0,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Date: metricscache.OnlyDate(date(2022, 8, 30)),
|
|
||||||
Amount: 1,
|
|
||||||
},
|
|
||||||
}, 1},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "no holes",
|
|
||||||
args: args{
|
|
||||||
rows: []database.InsertWorkspaceAgentStatParams{
|
|
||||||
statRow(zebra, dateH(2022, 8, 27, 0)),
|
|
||||||
statRow(zebra, dateH(2022, 8, 28, 0)),
|
|
||||||
statRow(zebra, dateH(2022, 8, 29, 0)),
|
|
||||||
},
|
|
||||||
},
|
|
||||||
tplWant: want{[]codersdk.DAUEntry{
|
|
||||||
{
|
|
||||||
Date: metricscache.OnlyDate(date(2022, 8, 27)),
|
|
||||||
Amount: 1,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Date: metricscache.OnlyDate(date(2022, 8, 28)),
|
|
||||||
Amount: 1,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Date: metricscache.OnlyDate(date(2022, 8, 29)),
|
|
||||||
Amount: 1,
|
|
||||||
},
|
|
||||||
}, 1},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "holes",
|
|
||||||
args: args{
|
|
||||||
rows: []database.InsertWorkspaceAgentStatParams{
|
|
||||||
statRow(zebra, dateH(2022, 1, 1, 0)),
|
|
||||||
statRow(tiger, dateH(2022, 1, 1, 0)),
|
|
||||||
statRow(zebra, dateH(2022, 1, 4, 0)),
|
|
||||||
statRow(zebra, dateH(2022, 1, 7, 0)),
|
|
||||||
statRow(tiger, dateH(2022, 1, 7, 0)),
|
|
||||||
},
|
|
||||||
},
|
|
||||||
tplWant: want{[]codersdk.DAUEntry{
|
|
||||||
{
|
|
||||||
Date: metricscache.OnlyDate(date(2022, 1, 1)),
|
|
||||||
Amount: 2,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Date: metricscache.OnlyDate(date(2022, 1, 2)),
|
|
||||||
Amount: 0,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Date: metricscache.OnlyDate(date(2022, 1, 3)),
|
|
||||||
Amount: 0,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Date: metricscache.OnlyDate(date(2022, 1, 4)),
|
|
||||||
Amount: 1,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Date: metricscache.OnlyDate(date(2022, 1, 5)),
|
|
||||||
Amount: 0,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Date: metricscache.OnlyDate(date(2022, 1, 6)),
|
|
||||||
Amount: 0,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Date: metricscache.OnlyDate(date(2022, 1, 7)),
|
|
||||||
Amount: 2,
|
|
||||||
},
|
|
||||||
}, 2},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "tzOffset",
|
|
||||||
tzOffset: 3,
|
|
||||||
args: args{
|
|
||||||
rows: []database.InsertWorkspaceAgentStatParams{
|
|
||||||
statRow(zebra, dateH(2022, 1, 2, 3)),
|
|
||||||
statRow(tiger, dateH(2022, 1, 2, 3)),
|
|
||||||
// With offset these should be in the previous day
|
|
||||||
statRow(zebra, dateH(2022, 1, 2, 0)),
|
|
||||||
statRow(tiger, dateH(2022, 1, 2, 0)),
|
|
||||||
},
|
|
||||||
},
|
|
||||||
tplWant: want{[]codersdk.DAUEntry{
|
|
||||||
{
|
|
||||||
Date: metricscache.OnlyDate(date(2022, 1, 2)),
|
|
||||||
Amount: 2,
|
|
||||||
},
|
|
||||||
}, 2},
|
|
||||||
dauWant: []codersdk.DAUEntry{
|
|
||||||
{
|
|
||||||
Date: metricscache.OnlyDate(date(2022, 1, 1)),
|
|
||||||
Amount: 2,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Date: metricscache.OnlyDate(date(2022, 1, 2)),
|
|
||||||
Amount: 2,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: "tzOffsetPreviousDay",
|
|
||||||
tzOffset: 6,
|
|
||||||
args: args{
|
|
||||||
rows: []database.InsertWorkspaceAgentStatParams{
|
|
||||||
statRow(zebra, dateH(2022, 1, 2, 1)),
|
|
||||||
statRow(tiger, dateH(2022, 1, 2, 1)),
|
|
||||||
statRow(zebra, dateH(2022, 1, 2, 0)),
|
|
||||||
statRow(tiger, dateH(2022, 1, 2, 0)),
|
|
||||||
},
|
|
||||||
},
|
|
||||||
dauWant: []codersdk.DAUEntry{
|
|
||||||
{
|
|
||||||
Date: metricscache.OnlyDate(date(2022, 1, 1)),
|
|
||||||
Amount: 2,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
tplWant: want{[]codersdk.DAUEntry{
|
|
||||||
{
|
|
||||||
Date: metricscache.OnlyDate(date(2022, 1, 2)),
|
|
||||||
Amount: 2,
|
|
||||||
},
|
|
||||||
}, 2},
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, tt := range tests {
|
|
||||||
tt := tt
|
|
||||||
t.Run(tt.name, func(t *testing.T) {
|
|
||||||
t.Parallel()
|
|
||||||
var (
|
|
||||||
db = dbmem.New()
|
|
||||||
cache = metricscache.New(db, slogtest.Make(t, nil), metricscache.Intervals{
|
|
||||||
TemplateDAUs: testutil.IntervalFast,
|
|
||||||
})
|
|
||||||
)
|
|
||||||
|
|
||||||
defer cache.Close()
|
|
||||||
|
|
||||||
template := dbgen.Template(t, db, database.Template{
|
|
||||||
Provisioner: database.ProvisionerTypeEcho,
|
|
||||||
})
|
|
||||||
|
|
||||||
for _, row := range tt.args.rows {
|
|
||||||
row.TemplateID = template.ID
|
|
||||||
row.ConnectionCount = 1
|
|
||||||
db.InsertWorkspaceAgentStat(context.Background(), row)
|
|
||||||
}
|
|
||||||
|
|
||||||
require.Eventuallyf(t, func() bool {
|
|
||||||
_, _, ok := cache.TemplateDAUs(template.ID, tt.tzOffset)
|
|
||||||
return ok
|
|
||||||
}, testutil.WaitShort, testutil.IntervalMedium,
|
|
||||||
"TemplateDAUs never populated",
|
|
||||||
)
|
|
||||||
|
|
||||||
gotUniqueUsers, ok := cache.TemplateUniqueUsers(template.ID)
|
|
||||||
require.True(t, ok)
|
|
||||||
|
|
||||||
if tt.dauWant != nil {
|
|
||||||
_, dauResponse, ok := cache.DeploymentDAUs(tt.tzOffset)
|
|
||||||
require.True(t, ok)
|
|
||||||
require.Equal(t, tt.dauWant, dauResponse.Entries)
|
|
||||||
}
|
|
||||||
|
|
||||||
offset, gotEntries, ok := cache.TemplateDAUs(template.ID, tt.tzOffset)
|
|
||||||
require.True(t, ok)
|
|
||||||
// Template only supports 0 offset.
|
|
||||||
require.Equal(t, 0, offset)
|
|
||||||
require.Equal(t, tt.tplWant.entries, gotEntries.Entries)
|
|
||||||
require.Equal(t, tt.tplWant.uniqueUsers, gotUniqueUsers)
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestCache_TemplateWorkspaceOwners(t *testing.T) {
|
func TestCache_TemplateWorkspaceOwners(t *testing.T) {
|
||||||
t.Parallel()
|
t.Parallel()
|
||||||
var ()
|
var ()
|
||||||
|
@ -261,7 +30,7 @@ func TestCache_TemplateWorkspaceOwners(t *testing.T) {
|
||||||
var (
|
var (
|
||||||
db = dbmem.New()
|
db = dbmem.New()
|
||||||
cache = metricscache.New(db, slogtest.Make(t, nil), metricscache.Intervals{
|
cache = metricscache.New(db, slogtest.Make(t, nil), metricscache.Intervals{
|
||||||
TemplateDAUs: testutil.IntervalFast,
|
TemplateBuildTimes: testutil.IntervalFast,
|
||||||
})
|
})
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -412,7 +181,7 @@ func TestCache_BuildTime(t *testing.T) {
|
||||||
var (
|
var (
|
||||||
db = dbmem.New()
|
db = dbmem.New()
|
||||||
cache = metricscache.New(db, slogtest.Make(t, nil), metricscache.Intervals{
|
cache = metricscache.New(db, slogtest.Make(t, nil), metricscache.Intervals{
|
||||||
TemplateDAUs: testutil.IntervalFast,
|
TemplateBuildTimes: testutil.IntervalFast,
|
||||||
})
|
})
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
|
@ -788,29 +788,9 @@ func (api *API) patchTemplateMeta(rw http.ResponseWriter, r *http.Request) {
|
||||||
// @Success 200 {object} codersdk.DAUsResponse
|
// @Success 200 {object} codersdk.DAUsResponse
|
||||||
// @Router /templates/{template}/daus [get]
|
// @Router /templates/{template}/daus [get]
|
||||||
func (api *API) templateDAUs(rw http.ResponseWriter, r *http.Request) {
|
func (api *API) templateDAUs(rw http.ResponseWriter, r *http.Request) {
|
||||||
ctx := r.Context()
|
|
||||||
template := httpmw.TemplateParam(r)
|
template := httpmw.TemplateParam(r)
|
||||||
|
|
||||||
vals := r.URL.Query()
|
api.returnDAUsInternal(rw, r, []uuid.UUID{template.ID})
|
||||||
p := httpapi.NewQueryParamParser()
|
|
||||||
tzOffset := p.Int(vals, 0, "tz_offset")
|
|
||||||
p.ErrorExcessParams(vals)
|
|
||||||
if len(p.Errors) > 0 {
|
|
||||||
httpapi.Write(ctx, rw, http.StatusBadRequest, codersdk.Response{
|
|
||||||
Message: "Query parameters have invalid values.",
|
|
||||||
Validations: p.Errors,
|
|
||||||
})
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
_, resp, _ := api.metricsCache.TemplateDAUs(template.ID, tzOffset)
|
|
||||||
if resp == nil || resp.Entries == nil {
|
|
||||||
httpapi.Write(ctx, rw, http.StatusOK, &codersdk.DAUsResponse{
|
|
||||||
Entries: []codersdk.DAUEntry{},
|
|
||||||
})
|
|
||||||
return
|
|
||||||
}
|
|
||||||
httpapi.Write(ctx, rw, http.StatusOK, resp)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// @Summary Get template examples by organization
|
// @Summary Get template examples by organization
|
||||||
|
|
Loading…
Reference in New Issue