commit 0d2efe4bcc9fe5f2722030ee39c8993c40134241 Author: Jonathan Bernard Date: Fri Feb 15 00:00:20 2019 -0600 Initial thoughts around API, DB layer. diff --git a/api/.api.txt.swp b/api/.api.txt.swp new file mode 100644 index 0000000..d7befcc Binary files /dev/null and b/api/.api.txt.swp differ diff --git a/api/.personal_measure_api.nimble.swp b/api/.personal_measure_api.nimble.swp new file mode 100644 index 0000000..57daad3 Binary files /dev/null and b/api/.personal_measure_api.nimble.swp differ diff --git a/api/Makefile b/api/Makefile new file mode 100644 index 0000000..0fae820 --- /dev/null +++ b/api/Makefile @@ -0,0 +1,25 @@ +PGSQL_CONTAINER_ID=`cat postgres.container.id` +DB_NAME="personal_measure" + +start-db: start-postgres + +postgres.container.id: + docker run --name postgres-$(DB_NAME) -e POSTGRES_PASSWORD=password -p 5500:5432 -d postgres > postgres.container.id + sleep 5 + PGPASSWORD=password psql -p 5500 -U postgres -h localhost -c "CREATE DATABASE $(DB_NAME);" + db_migrate up -c database-local.json + +start-postgres: postgres.container.id + docker start $(PGSQL_CONTAINER_ID) + db_migrate up -c database-local.json + +stop-postgres: postgres.container.id + docker stop $(PGSQL_CONTAINER_ID) + +delete-postgres-container: + -docker stop $(PGSQL_CONTAINER_ID) + docker container rm $(PGSQL_CONTAINER_ID) + rm postgres.container.id + +connect: + PGPASSWORD=password psql -p 5500 -U postgres -h localhost diff --git a/api/api.txt b/api/api.txt new file mode 100644 index 0000000..67ac8d3 --- /dev/null +++ b/api/api.txt @@ -0,0 +1,31 @@ +Personal Measure API +==================== + +### Measure + +☐ GET /measure Get a list of all defined measures for this user. +☐ POST /measure Create a new measure (post the definition). +☐ GET /measure/ Get the definition for a specific measure. +☐ DELETE /measure/ Delete a measure (and all values associated with it). + +### Values + +☐ GET / Get a list of values for a measure. +☐ POST / Add a new value for a measure. +☐ GET // Get the details for a specific value by id. +☐ PUT // Update a value by id. +☐ DELETE / Delete a value by id. + +### Auth + +☐ GET /auth-token Given a valid username/password combo, get an auth token. +☐ GET /user Given a valid auth token, return the user details. +☐ PSOT /app-token With a valid session, create a new app token. + +Legend +------ + +☐ TODO +☑ Done +☒ Will not do + diff --git a/api/database-local.json b/api/database-local.json new file mode 100644 index 0000000..b5d15f3 --- /dev/null +++ b/api/database-local.json @@ -0,0 +1,5 @@ +{ + "driver": "postgres", + "connectionString": "host=localhost port=5500 dbname=personal_measure user=postgres password=password", + "sqlDir": "src/main/sql/migrations" +} diff --git a/api/personal_measure_api b/api/personal_measure_api new file mode 100755 index 0000000..3b4c376 Binary files /dev/null and b/api/personal_measure_api differ diff --git a/api/personal_measure_api.nimble b/api/personal_measure_api.nimble new file mode 100644 index 0000000..5d73da2 --- /dev/null +++ b/api/personal_measure_api.nimble @@ -0,0 +1,14 @@ +# Package + +version = "0.1.0" +author = "Jonathan Bernard" +description = "JDB\'s Personal Measures API" +license = "MIT" +srcDir = "src/main/nim" +bin = @["personal_measure_api"] + + +# Dependencies + +requires @["nim >= 0.19.4", "bcrypt", "docopt >= 0.6.8", "isaac >= 0.1.3", "jester >= 0.4.1", "jwt", "tempfile", + "timeutils >= 0.4.0", "uuids >= 0.1.10" ] diff --git a/api/postgres.container.id b/api/postgres.container.id new file mode 100644 index 0000000..9d51ff2 --- /dev/null +++ b/api/postgres.container.id @@ -0,0 +1 @@ +7fcd40172b1207f53db40a9d4843ef5815588a742d8a334a767d07367281ff84 diff --git a/api/src/main/nim/.personal_measure_api.nim.swp b/api/src/main/nim/.personal_measure_api.nim.swp new file mode 100644 index 0000000..cb9542b Binary files /dev/null and b/api/src/main/nim/.personal_measure_api.nim.swp differ diff --git a/api/src/main/nim/personal_measure_api.nim b/api/src/main/nim/personal_measure_api.nim new file mode 100644 index 0000000..4166868 --- /dev/null +++ b/api/src/main/nim/personal_measure_api.nim @@ -0,0 +1,25 @@ +import asyncdispatch, bcrypt, docopt, jester, json, jwt + +import personal_measure_apipkg/db + +type + PersonalMeasureApiConfig = object + port*: int + pwdCost*: int + dbConnString*: string + +proc start*(cfg: StrawBossConfig): void = + + var stopFuture = newFuture[void]() + var workers: seq[Worker] = @[] + + settings: + port = Port(cfg.port) + appName = "/api" + + routes: + + get "/version": + resp($(%("strawboss v" & SB_VERSION)), JSON) + + diff --git a/api/src/main/nim/personal_measure_apipkg/.db.nim.swp b/api/src/main/nim/personal_measure_apipkg/.db.nim.swp new file mode 100644 index 0000000..8e6cc71 Binary files /dev/null and b/api/src/main/nim/personal_measure_apipkg/.db.nim.swp differ diff --git a/api/src/main/nim/personal_measure_apipkg/db b/api/src/main/nim/personal_measure_apipkg/db new file mode 100755 index 0000000..10a7594 Binary files /dev/null and b/api/src/main/nim/personal_measure_apipkg/db differ diff --git a/api/src/main/nim/personal_measure_apipkg/db.nim b/api/src/main/nim/personal_measure_apipkg/db.nim new file mode 100644 index 0000000..763b1d3 --- /dev/null +++ b/api/src/main/nim/personal_measure_apipkg/db.nim @@ -0,0 +1,170 @@ +import db_postgres, macros, options, postgres, sequtils, strutils, times, timeutils, uuids + +import nre except toSeq +from unicode import capitalize, toLower + +type + User* = object + id*: UUID + displayName*, email*, hashedPwd*: string + + ApiToken* = object + id*, userId: UUID + name*, hashedToken: string + expires: Option[DateTime] + + Measure* = object + id*, userId*: UUID + slug*, name*, description*, domainUnits*, rangeUnits*: string + domainSource*, rangeSource*: Option[string] + analysis*: seq[string] + + Value* = object + id*, measureId*: UUID + value: int + timestamp: DateTime + extData: string + + MutateClauses = object + columns*: seq[string] + placeholders*: seq[string] + values*: seq[string] + +let UPPERCASE_PATTERN = re"(.)(\p{Lu})" + +proc newMutateClauses(): MutateClauses = + return MutateClauses( + columns: @[], + placeholders: @[], + values: @[]) + +proc dbFormat(s: string): string = return s + +proc dbFormat(dt: DateTime): string = return dt.formatIso8601 + +proc dbFormat[T](list: seq[T]): string = + return "{" & list.mapIt("'" & dbFormat(it) & "'").join(",") & "}" + +proc dbFormat[T](item: T): string = return $item + +macro populateMutateClauses(t: typed, newRecord: bool, mc: var MutateClauses): untyped = + + # Must be working with an object. + let tTypeImpl = t.getTypeImpl + if not (tTypeImpl.typeKind == ntyObject): + error $t & " is not an object." + + result = newStmtList() + + # iterate over all the object's fields + for child in tTypeImpl[2].children: + + # ignore AST nodes that are not field definitions + if child.kind == nnkIdentDefs: + + # grab the field, it's string name, and it's type + let field = child[0] + let fieldType = child[1] + let fieldName = $field + + # we do not update the ID, but we do check: if we're creating a new + # record, we should not have an existing ID + if fieldName == "id": + result.add quote do: + if `newRecord` and not `t`.id.isZero: + raise newException( + AssertionError, + "Trying to create a new record, but the record already has an ID.") + + # if we're looking at an optional field, add logic to check for presence + elif fieldType.kind == nnkBracketExpr and + fieldType.len > 0 and + fieldType[0] == Option.getType: + + result.add quote do: + `mc`.columns.add(`fieldName`) + if `t`.`field`.isSome: + `mc`.placeholders.add("?") + `mc`.values.add(dbFormat(`t`.`field`.get)) + else: + `mc`.placeholders.add("NULL") + + # otherwise assume we can convert and go ahead. + else: + result.add quote do: + `mc`.columns.add(`fieldName`) + `mc`.placeholders.add("?") + `mc`.values.add(dbFormat(`t`.`field`)) + + #echo result.repr + +# TODO +#macro rowToRecord(recType: untyped, row: seq[string]): untyped +# echo recType + +macro recordName(rec: typed): string = + return $rec.getTypeInst + +proc tableNameForRecord[T](rec: T): string = + return recordName(rec).replace(UPPERCASE_PATTERN, "$1_$2").toLower() & "s" + +proc createRecord[T](db: DbConn, rec: T): T = + var mc = newMutateClauses() + populateMutateClauses(rec, true, mc) + + # Confusingly, getRow allows inserts and updates. We use it to get back the ID + # we want from the row. + let newIdStr = db.getValue(sql( + "INSERT INTO " & tableNameForRecord(rec) & + " (" & mc.columns.join(",") & ") " & + " VALUES (" & mc.placeholders.join(",") & ") " & + " RETURNING id"), mc.values) + + result = rec + result.id = parseUUID(newIdStr) + +proc updateRecord[T](db: DbConn, rec: T): bool = + var mc = newMutateClauses() + populateMutateClauses(rec, false, mc) + + let setClause = zip(mc.columns, mc.placeholders).mapIt(it.a & " = " it.b).join(',') + let numRowsUpdated = db.execAffectedRows(sql( + "UPDATE " & tableNameForRecord(rec) & + " SET " & setClause & + " WHERE id = ? "), mc.values.concat(@[rec.id])) + + return numRowsUpdated > 0; + +# TODO +#proc getRecord[T](db: DbConn, UUID id): T = + +macro listFieldNames(t: typed): untyped = + let tTypeImpl = t.getTypeImpl + + if not (tTypeImpl.typeKind == ntyObject): + error $t & " is not an object." + + var fieldNames: seq[tuple[n: string, t: string]] = @[] + for child in tTypeImpl[2].children: + if child.kind == nnkIdentDefs: + echo $child[1] + fieldNames.add((n: $child[0], t: $child[1])) + + result = newLit(fieldNames) + + +proc createUser*(db: DbConn, user: User): User = db.createRecord(user) +proc createApiToken*(db: DbConn, token: ApiToken): ApiToken = db.createRecord(token) +proc createMeasure*(db: DbConn, measure: Measure): Measure = db.createRecord(measure) +proc createValue*(db: DbConn, value: Value): Value = db.createRecord(value) + +#[ +when isMainModule: + + let u = User( + displayName: "Bob", + email: "bob@bobsco.com", + hashedPwd: "test") + + echo createRecord(nil, u) +]# diff --git a/api/src/main/sql/migrations/.20190214122514-initial-schema-up.sql.swp b/api/src/main/sql/migrations/.20190214122514-initial-schema-up.sql.swp new file mode 100644 index 0000000..1c878b9 Binary files /dev/null and b/api/src/main/sql/migrations/.20190214122514-initial-schema-up.sql.swp differ diff --git a/api/src/main/sql/migrations/20190214122514-initial-schema-down.sql b/api/src/main/sql/migrations/20190214122514-initial-schema-down.sql new file mode 100644 index 0000000..c31d3c6 --- /dev/null +++ b/api/src/main/sql/migrations/20190214122514-initial-schema-down.sql @@ -0,0 +1,5 @@ +-- DOWN script for initial-schema (20190214122514) +drop table if exists "values"; +drop table if exists "measures"; +drop table if exists "api_tokens"; +drop table if exists "users"; diff --git a/api/src/main/sql/migrations/20190214122514-initial-schema-up.sql b/api/src/main/sql/migrations/20190214122514-initial-schema-up.sql new file mode 100644 index 0000000..c4ad3f6 --- /dev/null +++ b/api/src/main/sql/migrations/20190214122514-initial-schema-up.sql @@ -0,0 +1,38 @@ +-- UP script for initial-schema (20190214122514) +create extension if not exists "uuid-ossp"; + +create table "users" ( + id uuid default uuid_generate_v4() primary key, + display_name varchar not null, + email varchar not null, + hashedpwd varchar not null +); + +create table "api_tokens" ( + id uuid default uuid_generate_v4() primary key, + user_id uuid not null references users (id) on delete cascade on update cascade, + name varchar not null, + expires timestamp with time zone default null, + hashedToken varchar not null +); + +create table "measures" ( + id uuid default uuid_generate_v4() primary key, + user_id uuid not null references users (id) on delete cascade on update cascade, + slug varchar not null, + name varchar not null, + description varchar not null default '', + domain_source varchar default null , + domain_units varchar not null default '', + range_source varchar default null, + range_units varchar not null default '', + analysis varchar[] not null default '{}' +); + +create table "values" ( + id uuid default uuid_generate_v4() primary key, + measure_id uuid not null references measures (id) on delete cascade on update cascade, + value integer not null, + "timestamp" timestamp not null default current_timestamp, + ext_data jsonb not null default '{}'::json +);