added
stringdate 2024-11-18 17:54:29
2024-11-19 03:43:44
| created
timestamp[s]date 1970-01-01 00:00:00
2025-02-21 19:39:40
| id
stringlengths 40
40
| int_score
int64 2
5
| metadata
dict | score
float64 2
4.75
| source
stringclasses 1
value | text
stringlengths 0
5.64M
|
---|---|---|---|---|---|---|---|
2024-11-18T18:08:22.422379+00:00 | 2021-11-09T14:53:37 | 4b6f272389e5d5147b40be6aafaaa88475f38766 | 2 | {
"blob_id": "4b6f272389e5d5147b40be6aafaaa88475f38766",
"branch_name": "refs/heads/master",
"committer_date": "2021-11-09T14:53:37",
"content_id": "fad04122b6fe626b393de7dec8352ab90a0195bd",
"detected_licenses": [
"Unlicense"
],
"directory_id": "b2dfbd930723b1405668e39f9b6a5edae59bc1b4",
"extension": "md",
"filename": "readme.md",
"fork_events_count": 0,
"gha_created_at": "2021-11-10T14:05:18",
"gha_event_created_at": "2021-11-10T14:05:19",
"gha_language": null,
"gha_license_id": "Unlicense",
"github_id": 426644819,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 255,
"license": "Unlicense",
"license_type": "permissive",
"path": "/www/docker/readme.md",
"provenance": "stack-edu-markdown-0000.json.gz:1",
"repo_name": "jfuerlinger/leocloud",
"revision_date": "2021-11-09T14:53:37",
"revision_id": "3e863f262be0e391d7803cf6713c7242e5494a10",
"snapshot_id": "76e52ee2cb3ef4680a0e861b6718565017878e8e",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/jfuerlinger/leocloud/3e863f262be0e391d7803cf6713c7242e5494a10/www/docker/readme.md",
"visit_date": "2023-08-28T15:29:08.583165"
} | 2.125 | stackv2 | # HTML Website
This folder contains the web project directly served by nginx.
## Sample
This is just a demo example to show how nginx serves html, css and javascript directly from
the volume and only redirect api calls to the application server.
|
2024-11-18T18:08:22.459006+00:00 | 2014-10-04T09:26:31 | dcb164ea6d7af91503e79d180443000222f5e1ac | 2 | {
"blob_id": "dcb164ea6d7af91503e79d180443000222f5e1ac",
"branch_name": "refs/heads/master",
"committer_date": "2014-10-04T09:26:31",
"content_id": "0cc87040943bd3346bdb1ca63d4e3b6433ef94aa",
"detected_licenses": [
"MIT"
],
"directory_id": "dd21bb9e5e71b03132f29e1d01b374e86677a8de",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": null,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 324,
"license": "MIT",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:2",
"repo_name": "pon/mock-rest-api",
"revision_date": "2014-10-04T09:26:31",
"revision_id": "022e66d17bfa9caa77400e6290a1632196875e42",
"snapshot_id": "5fdd0591de770a989b5af9aa5b83870f3391ba44",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/pon/mock-rest-api/022e66d17bfa9caa77400e6290a1632196875e42/README.md",
"visit_date": "2016-09-10T12:38:48.434617"
} | 2.15625 | stackv2 | mock-rest-api
=============
Example application to mock RESTful API calls
# Purpose
This is an imperfect API implementation. The purpose is to demostrate that you can mock the majority of the functionality of a RESTful JSON API using a relatively small number of lines of code.
# Usage
```
npm install
node index.js
```
|
2024-11-18T18:08:22.737924+00:00 | 2021-01-28T06:33:22 | f5ad76b2491cc0dc3fedae049cee805f6d60b897 | 2 | {
"blob_id": "f5ad76b2491cc0dc3fedae049cee805f6d60b897",
"branch_name": "refs/heads/master",
"committer_date": "2021-01-28T06:33:22",
"content_id": "6484be3a69f0f3e2792fdc318ef1091439c953ce",
"detected_licenses": [
"MIT"
],
"directory_id": "249ceeca8cd83e986cb838aeaf3cfa289c7c50e4",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": null,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 550,
"license": "MIT",
"license_type": "permissive",
"path": "/FundTransferApp/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:3",
"repo_name": "ihusnainalii/my-swift-journey",
"revision_date": "2021-01-28T06:33:22",
"revision_id": "72bb9fe115f94af0cad0d5454736e5c026dbc4f5",
"snapshot_id": "1c73103ad087a5a40f72e8b937da5cf7b992c81e",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/ihusnainalii/my-swift-journey/72bb9fe115f94af0cad0d5454736e5c026dbc4f5/FundTransferApp/README.md",
"visit_date": "2023-02-24T19:11:49.401255"
} | 2.03125 | stackv2 | # Fund Transfer App
> Build Fund Transfer application with SwiftUI.
![ |
2024-11-18T18:08:22.832718+00:00 | 2020-06-10T00:00:50 | 343d9a02f7b596d07737a6419b1bcd212a837544 | 2 | {
"blob_id": "343d9a02f7b596d07737a6419b1bcd212a837544",
"branch_name": "refs/heads/master",
"committer_date": "2020-06-10T00:00:50",
"content_id": "e82ec04d39b2640a72879987450f8e33b4929df5",
"detected_licenses": [
"Apache-2.0"
],
"directory_id": "243c2c4e34b9b00dad495e3d389dda37b3fe1885",
"extension": "md",
"filename": "2019-05-16.md",
"fork_events_count": 0,
"gha_created_at": "2018-03-28T14:10:39",
"gha_event_created_at": "2018-03-28T14:10:41",
"gha_language": null,
"gha_license_id": "Apache-2.0",
"github_id": 127150247,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 1274,
"license": "Apache-2.0",
"license_type": "permissive",
"path": "/en/Working_Group_People/2019-05-16.md",
"provenance": "stack-edu-markdown-0000.json.gz:4",
"repo_name": "petersilva/OS-Advisory_Conseil-SO",
"revision_date": "2020-06-10T00:00:50",
"revision_id": "70c19c74f77c26f288c52d4afeda019652f8e052",
"snapshot_id": "cb03e00b96b494a2cfcedeae485c499aec12807a",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/petersilva/OS-Advisory_Conseil-SO/70c19c74f77c26f288c52d4afeda019652f8e052/en/Working_Group_People/2019-05-16.md",
"visit_date": "2022-11-06T16:09:10.344794"
} | 2.203125 | stackv2 | # People Working Group Meeting
2019-05-16
## Participants
* Paul Bellows, Yellow Pencil
* Jim Cowie, ESDC
* Guillaume Charest, TBS
* Richard Seguin, Redhat
## Review Action items
* For communications, identify who the architects, Ops, and Devs people are in your department
* Speak to comms team about having a central landing page on the TBS website
* Guillaume to follow-up with SSC- enterprise architecture for help with ops
## Hackathon
* Link event: https://www.eventbrite.ca/e/hackathon-ogp-global-summit-2019-sommet-mondial-du-pgo-2019-tickets-61159685265
## Digital Academy
* Great experience but lots more awareness required to scale.
## Roundtable
* Richard: Need to kick start the working groups again, need to bring in new people.
* GCcollab posting on a group (which groups?)
* Twitter: whiteboard on amount of people using the platform
* Jim: Find a common problem to have multiple people work on it, multiple departments and private sector partners.
* Paul: tailored message for various stakeholders on Why open source: practitioners, executives, etc.
* Once we can find what types of messages, we can focus on where the groups are and channel it accordingly.
## Action items
* After summit: Guillaume to put KanBan up for each working groups
|
2024-11-18T18:08:22.954635+00:00 | 2017-04-03T22:33:16 | ae8aba086fb53eee07f8499249b22ee0a395ee76 | 2 | {
"blob_id": "ae8aba086fb53eee07f8499249b22ee0a395ee76",
"branch_name": "refs/heads/master",
"committer_date": "2017-04-03T22:33:16",
"content_id": "9514e2aac0c0dd44f56ee6ab2843cde2e580e523",
"detected_licenses": [
"MIT"
],
"directory_id": "6d8f3149df4d81410243793d9ba946407beebc06",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": 21862493,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 447,
"license": "MIT",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:5",
"repo_name": "bpapaspyros/RobSim",
"revision_date": "2017-04-03T22:33:16",
"revision_id": "b126e892910085fde4d14f50380b8db44fa7cd24",
"snapshot_id": "ed6f7e8ed739a3d813084fe4fe2733682ef6e14b",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/bpapaspyros/RobSim/b126e892910085fde4d14f50380b8db44fa7cd24/README.md",
"visit_date": "2021-01-23T14:51:32.490531"
} | 2.359375 | stackv2 | * To compile the program and produce the executables
> Installing Dependancies:
```
sudo apt-get install freeglut3-dev libxmu-dev libxi-dev libsoil-dev cmake
```
> Compiling:
```
mkdir build && cd build
cmake ..
make (-j #cores if available)
cd ../bin
```
> Running:
Without Graphics - Console Version
```
./Robsim
```
With Graphics
```
./RobsimGui
```
* There is a generated doxy for the code documentation, just run the script doxy.sh. |
2024-11-18T18:08:23.123812+00:00 | 2020-04-28T02:24:26 | 985e9ad70144e05ce41d462e04878ec8f7be9a3a | 3 | {
"blob_id": "985e9ad70144e05ce41d462e04878ec8f7be9a3a",
"branch_name": "refs/heads/master",
"committer_date": "2020-04-28T02:24:26",
"content_id": "7bd40941d89026f44fbd50c2692303daf5c35925",
"detected_licenses": [
"MIT"
],
"directory_id": "d96b4cc20fd4ec43329394bab12dd174c4baabbc",
"extension": "md",
"filename": "README.md",
"fork_events_count": 2,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": 157544833,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 555,
"license": "MIT",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:6",
"repo_name": "pangyuteng/closed-contour-segmentation-using-graph-theory",
"revision_date": "2020-04-28T02:24:26",
"revision_id": "e9c8885dd98ccc250ab3fe598c1c83d7d3a76058",
"snapshot_id": "1593c61600c5824e3ea1fdd0eb6fb59027fdc3c6",
"src_encoding": "UTF-8",
"star_events_count": 3,
"url": "https://raw.githubusercontent.com/pangyuteng/closed-contour-segmentation-using-graph-theory/e9c8885dd98ccc250ab3fe598c1c83d7d3a76058/README.md",
"visit_date": "2021-08-04T13:47:07.537682"
} | 2.640625 | stackv2 | ### Closed-contour segmentation using graph theory and dynamic programming
#### Usage.
see `demo.ipynb` for usage.
# get contour coordinates and mask from demo image
px,py,mask = get_closed_contour(testimg)

#### Reference.
Chiu SJ, Toth CA, Rickman CB, Izatt JA, Farsiu S. Automatic segmentation of closed-contour features in ophthalmic images using graph theory and dynamic programming. Biomedical optics express. 2012 May 1;3(5):1127-40.
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3342188/?report=classic |
2024-11-18T18:08:23.161424+00:00 | 2020-05-10T01:35:38 | e72da24ea0fa5a0188c2be269213832f808769d4 | 2 | {
"blob_id": "e72da24ea0fa5a0188c2be269213832f808769d4",
"branch_name": "refs/heads/master",
"committer_date": "2020-05-10T01:35:38",
"content_id": "a86dd9e9d079f9d7ada9f62a9f7e76def0231be4",
"detected_licenses": [
"MIT"
],
"directory_id": "5533e1a838cf0ce3fa699d4dc6478f356f097f88",
"extension": "md",
"filename": "honey.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": null,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 90660,
"license": "MIT",
"license_type": "permissive",
"path": "/index/h/honey.md",
"provenance": "stack-edu-markdown-0000.json.gz:7",
"repo_name": "otac0n/recipes",
"revision_date": "2020-05-10T01:35:38",
"revision_id": "48268785a13d598a87de2e75c525056d00202e54",
"snapshot_id": "0d6d7611086a7b6f41940044faf4dc7eabc7e61c",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/otac0n/recipes/48268785a13d598a87de2e75c525056d00202e54/index/h/honey.md",
"visit_date": "2023-08-22T19:10:19.399625"
} | 2.484375 | stackv2 | # honey
* [Coffee And Turron Ice Cream Torte With Honey Brandy Oranges](../../index/c/coffee-and-turron-ice-cream-torte-with-honey-brandy-oranges-107168.json)
* [Cornish Game Hens With Dried Fruit And Honey](../../index/c/cornish-game-hens-with-dried-fruit-and-honey-105525.json)
* [Cornmeal Pancakes With Honey Pecan Butter](../../index/c/cornmeal-pancakes-with-honey-pecan-butter-108554.json)
* [Cream Tart With Oranges Honey And Toasted Almond Crust](../../index/c/cream-tart-with-oranges-honey-and-toasted-almond-crust-357275.json)
* [Creme Fraiche Cheesecake With Honey Rum Roasted Pineapple](../../index/c/creme-fraiche-cheesecake-with-honey-rum-roasted-pineapple-109089.json)
* [Crispy Lamb Spare Ribs With Honey Creme Fraiche And Chopped Pea Salad](../../index/c/crispy-lamb-spare-ribs-with-honey-creme-fraiche-and-chopped-pea-salad-241226.json)
* [Dark Ginger Rye Cake With Yogurt And Honey](../../index/d/dark-ginger-rye-cake-with-yogurt-and-honey.json)
* [Date And Walnut Phyllo Rolls With Greek Yogurt And Honey](../../index/d/date-and-walnut-phyllo-rolls-with-greek-yogurt-and-honey-233722.json)
* [Duck Breasts With Orange Honey And Tea Sauce](../../index/d/duck-breasts-with-orange-honey-and-tea-sauce-103268.json)
* [Duck With Honey](../../index/d/duck-with-honey-101622.json)
* [Duck With Honey Peppercorn Sauce](../../index/d/duck-with-honey-peppercorn-sauce-4690.json)
* [Earl Grey Tea Madeleines With Honey](../../index/e/earl-grey-tea-madeleines-with-honey-231609.json)
* [Feta With Pepper Honey](../../index/f/feta-with-pepper-honey-100908.json)
* [Feta With Rosemary Pepper Honey](../../index/f/feta-with-rosemary-pepper-honey-13096.json)
* [Figs And Dried Cherries Poached In Vanilla Honey Syrup With Pistachios](../../index/f/figs-and-dried-cherries-poached-in-vanilla-honey-syrup-with-pistachios-107510.json)
* [Figs With Honey Orange Mascarpone And Pistachio Brittle](../../index/f/figs-with-honey-orange-mascarpone-and-pistachio-brittle-361232.json)
* [Fluffy Cathead Biscuits With Honey Butter](../../index/f/fluffy-cathead-biscuits-with-honey-butter.json)
* [Fresh Fruit Salad With Honey Mint And Lime Syrup](../../index/f/fresh-fruit-salad-with-honey-mint-and-lime-syrup-15546.json)
* [Fresh Fruit With Honey Yogurt Dip](../../index/f/fresh-fruit-with-honey-yogurt-dip-5457.json)
* [Fried Chicken And Biscuits With Pink Peppercorn Honey Butter](../../index/f/fried-chicken-and-biscuits-with-pink-peppercorn-honey-butter-354874.json)
* [Fried Eggplant With Honey Mint And Sesame Seeds](../../index/f/fried-eggplant-with-honey-mint-and-sesame-seeds-107127.json)
* [Fried Pastry Spirals With Honey Sesame And Walnuts](../../index/f/fried-pastry-spirals-with-honey-sesame-and-walnuts-108130.json)
* [Frozen Honey Mousse With Tart Cherry Cookies](../../index/f/frozen-honey-mousse-with-tart-cherry-cookies-364675.json)
* [Fruit Salad With Honey Lime Dressing](../../index/f/fruit-salad-with-honey-lime-dressing-106941.json)
* [Gala Apple Newtons With Honey Cream](../../index/g/gala-apple-newtons-with-honey-cream-108699.json)
* [Game Hens With Honey Vinegar And Spice Glaze](../../index/g/game-hens-with-honey-vinegar-and-spice-glaze-1886.json)
* [Ginger And Honey Baby Back Ribs](../../index/g/ginger-and-honey-baby-back-ribs-353813.json)
* [Ginger Honey Lemonade](../../index/g/ginger-honey-lemonade-201200.json)
* [Goat Cheese And Honey Blancmange](../../index/g/goat-cheese-and-honey-blancmange-240232.json)
* [Goat Cheese Toasts With Walnuts Honey Thyme](../../index/g/goat-cheese-toasts-with-walnuts-honey-thyme-51166200.json)
* [Greek Honey And Anise Twists](../../index/g/greek-honey-and-anise-twists-3051.json)
* [Grilled Corn With Honey Ginger Barbecue Sauce](../../index/g/grilled-corn-with-honey-ginger-barbecue-sauce-51108010.json)
* [Grilled Lamb Chops With Red Wine Garlic And Honey Glaze](../../index/g/grilled-lamb-chops-with-red-wine-garlic-and-honey-glaze-5610.json)
* [Grilled Lemon Honey Chicken Breasts](../../index/g/grilled-lemon-honey-chicken-breasts-101958.json)
* [Grilled Nectarines With Honey Balsamic Glaze](../../index/g/grilled-nectarines-with-honey-balsamic-glaze-106651.json)
* [Grilled Pineapple And Banana Sauce With Honey And Lime](../../index/g/grilled-pineapple-and-banana-sauce-with-honey-and-lime-239044.json)
* [Grilled Quail Salad With Honey Dijon Dressing](../../index/g/grilled-quail-salad-with-honey-dijon-dressing-239843.json)
* [Grilled Shrimp With Honey Ginger Barbecue Sauce](../../index/g/grilled-shrimp-with-honey-ginger-barbecue-sauce-51108030.json)
* [Grilled Skewered Scallops And Apricots With Honey Mustard Dressing](../../index/g/grilled-skewered-scallops-and-apricots-with-honey-mustard-dressing-235338.json)
* [Ham And Munster Buns With Honey Mustard](../../index/h/ham-and-munster-buns-with-honey-mustard-10545.json)
* [Ham Stuffed Chicken Legs With Sweet Pepper Honey And Onion Compote](../../index/h/ham-stuffed-chicken-legs-with-sweet-pepper-honey-and-onion-compote-1964.json)
* [High Honey Sriracha Popcorn](../../index/h/high-honey-sriracha-popcorn.json)
* [Hoisin And Honey Glazed Pork Chops](../../index/h/hoisin-and-honey-glazed-pork-chops-109015.json)
* [Hoisin And Honey Pork Riblets](../../index/h/hoisin-and-honey-pork-riblets-12141.json)
* [Honey Almond And Date Ice Cream Sauce](../../index/h/honey-almond-and-date-ice-cream-sauce-10391.json)
* [Honey Almond Granola](../../index/h/honey-almond-granola-350875.json)
* [Honey And Soy Glazed Salmon](../../index/h/honey-and-soy-glazed-salmon-14110.json)
* [Honey And Spice Cranberry Sauce](../../index/h/honey-and-spice-cranberry-sauce-100322.json)
* [Honey And Spice Loaf Cake](../../index/h/honey-and-spice-loaf-cake-102698.json)
* [Honey And Thyme Ice Cream With Candied Thyme](../../index/h/honey-and-thyme-ice-cream-with-candied-thyme-236890.json)
* [Honey And Vinegar Braised Short Ribs With Spinach](../../index/h/honey-and-vinegar-braised-short-ribs-with-spinach-102858.json)
* [Honey Anise And Almond Biscotti](../../index/h/honey-anise-and-almond-biscotti-103158.json)
* [Honey Barbequed Salmon With Mache](../../index/h/honey-barbequed-salmon-with-mache-12667.json)
* [Honey Bourbon Toddy](../../index/h/honey-bourbon-toddy-368975.json)
* [Honey Brandy Oranges](../../index/h/honey-brandy-oranges-107171.json)
* [Honey Brined Turkey With Giblet Cream Gravy](../../index/h/honey-brined-turkey-with-giblet-cream-gravy-102478.json)
* [Honey Butter French Bread](../../index/h/honey-butter-french-bread-102470.json)
* [Honey Caramel Tart With Apricots And Almonds](../../index/h/honey-caramel-tart-with-apricots-and-almonds-237924.json)
* [Honey Cardamom Custards With Strawberry Orange Compote](../../index/h/honey-cardamom-custards-with-strawberry-orange-compote-1580.json)
* [Honey Cheese Pie With Strawberry Compote](../../index/h/honey-cheese-pie-with-strawberry-compote-1705.json)
* [Honey Cinnamon Ice Cream](../../index/h/honey-cinnamon-ice-cream-364329.json)
* [Honey Citrus Chicken](../../index/h/honey-citrus-chicken-2280.json)
* [Honey Custard In Phyllo Pastry](../../index/h/honey-custard-in-phyllo-pastry-1070.json)
* [Honey Date And Pecan Tart](../../index/h/honey-date-and-pecan-tart-241628.json)
* [Honey Date Nut Breads](../../index/h/honey-date-nut-breads-11080.json)
* [Honey Dew And Sake Granita](../../index/h/honey-dew-and-sake-granita-12376.json)
* [Honey Eggnog](../../index/h/honey-eggnog-200915.json)
* [Honey Ginger Barbecue Sauce](../../index/h/honey-ginger-barbecue-sauce-51108020.json)
* [Honey Ginger Pineapple Crepes](../../index/h/honey-ginger-pineapple-crepes-103208.json)
* [Honey Ginger Salmon](../../index/h/honey-ginger-salmon-5908.json)
* [Honey Gingered Pork Tenderloins](../../index/h/honey-gingered-pork-tenderloins-15277.json)
* [Honey Glazed Baked Apples Filled With Dried Fruit](../../index/h/honey-glazed-baked-apples-filled-with-dried-fruit-102520.json)
* [Honey Glazed Butternut Squash](../../index/h/honey-glazed-butternut-squash-4580.json)
* [Honey Glazed Carrots With Carrot Top Gremolata](../../index/h/honey-glazed-carrots-with-carrot-top-gremolata.json)
* [Honey Glazed Doughnuts With Raisins And Pine Nuts](../../index/h/honey-glazed-doughnuts-with-raisins-and-pine-nuts-236796.json)
* [Honey Glazed Duck With Fig And Pistachio Red Quinoa](../../index/h/honey-glazed-duck-with-fig-and-pistachio-red-quinoa-51125200.json)
* [Honey Glazed Grilled Pineapple](../../index/h/honey-glazed-grilled-pineapple-108252.json)
* [Honey Glazed Lemon Chicken](../../index/h/honey-glazed-lemon-chicken-109394.json)
* [Honey Glazed Oven Roasted Sweet Potato Wedges](../../index/h/honey-glazed-oven-roasted-sweet-potato-wedges-236503.json)
* [Honey Glazed Peach Tart With Mascarpone Cream](../../index/h/honey-glazed-peach-tart-with-mascarpone-cream-108473.json)
* [Honey Glazed Rack Of Lamb](../../index/h/honey-glazed-rack-of-lamb-4392.json)
* [Honey Glazed Vegetables](../../index/h/honey-glazed-vegetables.json)
* [Honey Hazelnut Financiers](../../index/h/honey-hazelnut-financiers-51234850.json)
* [Honey Lemon Buttermilk Dressing](../../index/h/honey-lemon-buttermilk-dressing-12083.json)
* [Honey Lemon Custard With Fruit](../../index/h/honey-lemon-custard-with-fruit-51104290.json)
* [Honey Mustard And Rosemary Pork Roast](../../index/h/honey-mustard-and-rosemary-pork-roast-2130.json)
* [Honey Mustard Chicken Sausage Kebabs](../../index/h/honey-mustard-chicken-sausage-kebabs-353417.json)
* [Honey Mustard Dressing](../../index/h/honey-mustard-dressing-11578.json)
* [Honey Mustard Dressing And Mixed Greens](../../index/h/honey-mustard-dressing-and-mixed-greens-104167.json)
* [Honey Mustard Glazed Shrimp And Scallops](../../index/h/honey-mustard-glazed-shrimp-and-scallops-238009.json)
* [Honey Mustard Sauce](../../index/h/honey-mustard-sauce-109036.json)
* [Honey Mustard Sauce](../../index/h/honey-mustard-sauce-5442.json)
* [Honey Nut Squares](../../index/h/honey-nut-squares-107829.json)
* [Honey Oat Bread](../../index/h/honey-oat-bread-105775.json)
* [Honey Orange Butternut Squash](../../index/h/honey-orange-butternut-squash-10650.json)
* [Honey Orange Madeleines](../../index/h/honey-orange-madeleines-1807.json)
* [Honey Pan Dulce With Nuts And Dried Fruit](../../index/h/honey-pan-dulce-with-nuts-and-dried-fruit-236470.json)
* [Honey Peach Tart With Vanilla Ice Cream](../../index/h/honey-peach-tart-with-vanilla-ice-cream-2300.json)
* [Honey Poached Pears With Mascarpone](../../index/h/honey-poached-pears-with-mascarpone-100535.json)
* [Honey Roasted Carrots With Tahini Yogurt](../../index/h/honey-roasted-carrots-with-tahini-yogurt-51254840.json)
* [Honey Roasted Cherry Tomatoes](../../index/h/honey-roasted-cherry-tomatoes-51174600.json)
* [Honey Roasted Chicken With Lemon And Tarragon](../../index/h/honey-roasted-chicken-with-lemon-and-tarragon-108489.json)
* [Honey Roasted Ham Or Turkey With Dried Cherry Relish](../../index/h/honey-roasted-ham-or-turkey-with-dried-cherry-relish-5774.json)
* [Honey Roasted Lamb With Arugula And Pine Nut Salad](../../index/h/honey-roasted-lamb-with-arugula-and-pine-nut-salad-106475.json)
* [Honey Roasted Onion Tart](../../index/h/honey-roasted-onion-tart-363711.json)
* [Honey Roasted Pear Salad With Thyme Verjus Dressing](../../index/h/honey-roasted-pear-salad-with-thyme-verjus-dressing-232870.json)
* [Honey Rosemary Ice Cream](../../index/h/honey-rosemary-ice-cream-363713.json)
* [Honey Rye](../../index/h/honey-rye-367137.json)
* [Honey Tree Bark Biscuits](../../index/h/honey-tree-bark-biscuits-243193.json)
* [Honey Turmeric Pork With Beet And Carrot Salad](../../index/h/honey-turmeric-pork-with-beet-and-carrot-salad-56389766.json)
* [Honey Vanilla Poached Apricots](../../index/h/honey-vanilla-poached-apricots-12387.json)
* [Honey Vinegar Leg Of Lamb With Fennel And Carrots](../../index/h/honey-vinegar-leg-of-lamb-with-fennel-and-carrots-51193690.json)
* [Honey Yogurt](../../index/h/honey-yogurt-354340.json)
* [Jalapeno Honey Mustard Pork Kebabs](../../index/j/jalapeno-honey-mustard-pork-kebabs-10142.json)
* [Just As Sweet As Honey Cake](../../index/j/just-as-sweet-as-honey-cake-369050.json)
* [Lavender Honey And Yogurt Pie](../../index/l/lavender-honey-and-yogurt-pie-238748.json)
* [Lavender Honey Tea Bread](../../index/l/lavender-honey-tea-bread-233911.json)
* [Lemon Honey Almond Cake](../../index/l/lemon-honey-almond-cake-101412.json)
* [Lemon Honey And Walnut Tart](../../index/l/lemon-honey-and-walnut-tart-1473.json)
* [Lemon Honey Tart With Salted Shortbread Crust](../../index/l/lemon-honey-tart-with-salted-shortbread-crust-51140700.json)
* [Lime And Honey Glazed Salmon With Basmati And Broccolini](../../index/l/lime-and-honey-glazed-salmon-with-basmati-and-broccolini-355189.json)
* [Low Fat Yogurt Honey Mint Dressing](../../index/l/low-fat-yogurt-honey-mint-dressing-12219.json)
* [Madeleines With Lavender Honey](../../index/m/madeleines-with-lavender-honey-357229.json)
* [Maltese Chicken With Red Wine And Honey Gravy](../../index/m/maltese-chicken-with-red-wine-and-honey-gravy-106485.json)
* [Marjoram Honey Vinaigrette](../../index/m/marjoram-honey-vinaigrette-240032.json)
* [Mascarpone Tart With Honey Oranges And Pistachios](../../index/m/mascarpone-tart-with-honey-oranges-and-pistachios-237299.json)
* [Mexican Chopped Salad With Honey Lime Dressing](../../index/m/mexican-chopped-salad-with-honey-lime-dressing-230154.json)
* [Mixed Greens With Honey Mustard Dressing](../../index/m/mixed-greens-with-honey-mustard-dressing-11398.json)
* [Mixed Greens With Sheeps Milk Cheese And Honey Vinaigrette](../../index/m/mixed-greens-with-sheeps-milk-cheese-and-honey-vinaigrette-236189.json)
* [Molasses Honey Ginger Cookies](../../index/m/molasses-honey-ginger-cookies-106791.json)
* [Nectarine Tarts With Honey Ice Cream](../../index/n/nectarine-tarts-with-honey-ice-cream-2330.json)
* [Nectarines Strawberries And Melon In Orange Honey Spiced Syrup](../../index/n/nectarines-strawberries-and-melon-in-orange-honey-spiced-syrup-5411.json)
* [Oat And Dried Fruit Cereal With Honey And Yogurt](../../index/o/oat-and-dried-fruit-cereal-with-honey-and-yogurt-4477.json)
* [Old Fashioned Oatmeal With Apples Raisins And Honey Toasted Walnuts](../../index/o/old-fashioned-oatmeal-with-apples-raisins-and-honey-toasted-walnuts-3116.json)
* [Orange Blossom Honey Madeleines](../../index/o/orange-blossom-honey-madeleines-367764.json)
* [Orange Honey Sauce](../../index/o/orange-honey-sauce-103922.json)
* [Orange Honey Syrup](../../index/o/orange-honey-syrup-10414.json)
* [Orange Roasted Baby Carrots With Honey](../../index/o/orange-roasted-baby-carrots-with-honey-237912.json)
* [Orange Spiced Rye Honey Cake Gefilte Manifesto Recipe](../../index/o/orange-spiced-rye-honey-cake-gefilte-manifesto-recipe.json)
* [Oranges With Pomegranate Molasses And Honey](../../index/o/oranges-with-pomegranate-molasses-and-honey-350797.json)
* [Oven Poached Pears With Lemon Honey And Nutmeg](../../index/o/oven-poached-pears-with-lemon-honey-and-nutmeg-5160.json)
* [Passover Honey Nut Cake In Soaking Syrup](../../index/p/passover-honey-nut-cake-in-soaking-syrup-109151.json)
* [Pastry Twists With Spiced Sugar Honey Glaze](../../index/p/pastry-twists-with-spiced-sugar-honey-glaze-358569.json)
* [Pear Cake With Lemon Honey Cream Cheese Frosting](../../index/p/pear-cake-with-lemon-honey-cream-cheese-frosting-104048.json)
* [Pear Crostata With Figs And Honey](../../index/p/pear-crostata-with-figs-and-honey-361795.json)
* [Pears In Honey And Pine Nut Caramel With Artisanal Cheese](../../index/p/pears-in-honey-and-pine-nut-caramel-with-artisanal-cheese-241350.json)
* [Pears With Herbed Ricotta And Honey](../../index/p/pears-with-herbed-ricotta-and-honey-232639.json)
* [Pecan Biscuits With Orange Honey Glaze](../../index/p/pecan-biscuits-with-orange-honey-glaze-5014.json)
* [Pecan Waffles With Sauteed Bananas And Cinnamon Honey](../../index/p/pecan-waffles-with-sauteed-bananas-and-cinnamon-honey-363321.json)
* [Phyllo Nests With Strawberries And Honey](../../index/p/phyllo-nests-with-strawberries-and-honey-108131.json)
* [Pine Nut Honey Ice Cream](../../index/p/pine-nut-honey-ice-cream-105233.json)
* [Plantain Flatbread With Poached Egg And Honey Drizzle](../../index/p/plantain-flatbread-with-poached-egg-and-honey-drizzle.json)
* [Plum Fennel Salad With Honey Ginger Dressing](../../index/p/plum-fennel-salad-with-honey-ginger-dressing-56389860.json)
* [Plum Tarts With Honey And Black Pepper](../../index/p/plum-tarts-with-honey-and-black-pepper-51239610.json)
* [Pork Kebabs Marinated In Honey Rosemary And Orange](../../index/p/pork-kebabs-marinated-in-honey-rosemary-and-orange-106479.json)
* [Port And Honey Poached Pears With Lemon Curd Mousse](../../index/p/port-and-honey-poached-pears-with-lemon-curd-mousse-12712.json)
* [Profiteroles With Honey Lavender Ice Cream](../../index/p/profiteroles-with-honey-lavender-ice-cream-103082.json)
* [Puff Pastry Apple Tarts Glazed With Honey](../../index/p/puff-pastry-apple-tarts-glazed-with-honey-107174.json)
* [Puff Pastry Baskets With Honey Mousse And Assorted Fruit](../../index/p/puff-pastry-baskets-with-honey-mousse-and-assorted-fruit-1904.json)
* [Red Chili And Honey Glazed Turkey With Ancho Pan Gravy](../../index/r/red-chili-and-honey-glazed-turkey-with-ancho-pan-gravy-104309.json)
* [Rice Pudding With Fresh Pears And Honey](../../index/r/rice-pudding-with-fresh-pears-and-honey-51246410.json)
* [Roast Chicken With Rosemary Lemon And Honey](../../index/r/roast-chicken-with-rosemary-lemon-and-honey-395509.json)
* [Roast Chicken With Saffron Hazelnuts And Honey](../../index/r/roast-chicken-with-saffron-hazelnuts-and-honey-51184640.json)
* [Roasted Apricots With Honey Vanilla Creme Fraiche](../../index/r/roasted-apricots-with-honey-vanilla-creme-fraiche-359249.json)
* [Roasted Honey Glazed Spareribs](../../index/r/roasted-honey-glazed-spareribs-100570.json)
* [Roasted Hot Honey Shrimp With Bok Choy And Kimchi Rice](../../index/r/roasted-hot-honey-shrimp-with-bok-choy-and-kimchi-rice-51261050.json)
* [Roasted Parsnips Turnips And Rutabagas With Ancho Spiced Honey Glaze](../../index/r/roasted-parsnips-turnips-and-rutabagas-with-ancho-spiced-honey-glaze-361291.json)
* [Roasted Pineapple With Honey And Pistachios](../../index/r/roasted-pineapple-with-honey-and-pistachios-394677.json)
* [Roasted Spiced Chicken With Cinnamon And Honey Glazed Sweet Potatoes](../../index/r/roasted-spiced-chicken-with-cinnamon-and-honey-glazed-sweet-potatoes-234808.json)
* [Roasted Sweet Potatoes With Honey Glaze](../../index/r/roasted-sweet-potatoes-with-honey-glaze-104728.json)
* [Rustic Apple Tart With Honey Dates And Nuts](../../index/r/rustic-apple-tart-with-honey-dates-and-nuts-108599.json)
* [Sage And Honey Skillet Cornbread](../../index/s/sage-and-honey-skillet-cornbread-240376.json)
* [Salmon Glazed With Honey And Mustard](../../index/s/salmon-glazed-with-honey-and-mustard-102649.json)
* [Salmon Glazed With Honey And Mustard](../../index/s/salmon-glazed-with-honey-and-mustard-350653.json)
* [Salmon With Soy Honey And Wasabi Sauces](../../index/s/salmon-with-soy-honey-and-wasabi-sauces-105026.json)
* [Scottish Salmon With Shallot Truffle Honey Glaze Lump Crab And Green Apple Risotto And Quince Jam](../../index/s/scottish-salmon-with-shallot-truffle-honey-glaze-lump-crab-and-green-apple-risotto-and-quince-jam-363590.json)
* [Seared Duck Breast With Chili Honey Ginger Glaze](../../index/s/seared-duck-breast-with-chili-honey-ginger-glaze-102303.json)
* [Seared Scallops With Spicy Honey Citrus Glaze On Celery Salad](../../index/s/seared-scallops-with-spicy-honey-citrus-glaze-on-celery-salad-239848.json)
* [Skewered Grilled Fruit With Minted Yogurt Honey Sauce](../../index/s/skewered-grilled-fruit-with-minted-yogurt-honey-sauce-105250.json)
* [Skewered Honey Glazed Scallops With Peach Salsa](../../index/s/skewered-honey-glazed-scallops-with-peach-salsa-109692.json)
* [Smoked Baby Back Ribs With Hoisin Honey Glaze](../../index/s/smoked-baby-back-ribs-with-hoisin-honey-glaze-105305.json)
* [Spiced Fresh Orange And Honey Sorbet](../../index/s/spiced-fresh-orange-and-honey-sorbet-241111.json)
* [Spiced Honey Cake With Cream Cheese Frosting](../../index/s/spiced-honey-cake-with-cream-cheese-frosting-56389999.json)
* [Spicy Honey Glazed Parsnips](../../index/s/spicy-honey-glazed-parsnips-51223670.json)
* [Spicy Honey Mustard Sauce](../../index/s/spicy-honey-mustard-sauce-56389343.json)
* [Strawberry Panna Cotta With Strawberries And Honey](../../index/s/strawberry-panna-cotta-with-strawberries-and-honey-103528.json)
* [Sweet Honey And Saffron Challah](../../index/s/sweet-honey-and-saffron-challah-361129.json)
* [Sweet Potato And Poblano Salad With Honey And Rosemary](../../index/s/sweet-potato-and-poblano-salad-with-honey-and-rosemary-353851.json)
* [Sweet Potato Biscuits With Ham Mustard And Honey](../../index/s/sweet-potato-biscuits-with-ham-mustard-and-honey-356312.json)
* [Sweet Potato Pancakes With Honey Cinnamon Butter](../../index/s/sweet-potato-pancakes-with-honey-cinnamon-butter-239216.json)
* [Sweet Potatoes And Greens With Honey Mustard Glaze](../../index/s/sweet-potatoes-and-greens-with-honey-mustard-glaze-2605.json)
* [Swiss Honey Walnut Tart](../../index/s/swiss-honey-walnut-tart-4663.json)
* [Tamarind Honey Lamb Kebabs On Mashed Yams](../../index/t/tamarind-honey-lamb-kebabs-on-mashed-yams-231084.json)
* [Tangerine Honey Flan With Grapefruit Segments](../../index/t/tangerine-honey-flan-with-grapefruit-segments-234825.json)
* [Tea And Honey Crisps](../../index/t/tea-and-honey-crisps-236667.json)
* [Tricolor Salad With Honey Cumin Dressing](../../index/t/tricolor-salad-with-honey-cumin-dressing-1738.json)
* [Turkey Cheddar Sandwiches With Honey Mustard](../../index/t/turkey-cheddar-sandwiches-with-honey-mustard-101442.json)
* [Twice Roasted Sweet Potatoes With Hot Honey](../../index/t/twice-roasted-sweet-potatoes-with-hot-honey.json)
* [Upside Down Honey Cheesecakes](../../index/u/upside-down-honey-cheesecakes-237917.json)
* [Vanilla Bean Roasted Figs With Wildflower Honey Vanilla Ice Cream](../../index/v/vanilla-bean-roasted-figs-with-wildflower-honey-vanilla-ice-cream-105862.json)
* [Vidalia Onion Tart With Bacon Local Honey And Fresh Thyme](../../index/v/vidalia-onion-tart-with-bacon-local-honey-and-fresh-thyme-354950.json)
* [Walnut Apricot Cake In Honey Syrup](../../index/w/walnut-apricot-cake-in-honey-syrup-106498.json)
* [Walnut Thyme Honey](../../index/w/walnut-thyme-honey-368990.json)
* [Warm Goat Cheese And Carrot Salad With Marjoram Honey Vinaigrette](../../index/w/warm-goat-cheese-and-carrot-salad-with-marjoram-honey-vinaigrette-240031.json)
* [Warm Goat Cheese Toasts With Rosemary Walnuts And Honey](../../index/w/warm-goat-cheese-toasts-with-rosemary-walnuts-and-honey-108799.json)
* [Whipped Ricotta With Honey And Mixed Berries](../../index/w/whipped-ricotta-with-honey-and-mixed-berries-105374.json)
* [White Sweet Potatoes With Mirin And Honey](../../index/w/white-sweet-potatoes-with-mirin-and-honey-388730.json)
* [Wildflower Honey Semifreddo With Honey Sesame Wafers](../../index/w/wildflower-honey-semifreddo-with-honey-sesame-wafers-232643.json)
* [Wildflower Honey Semifreddo With Honey Sesame Wafers](../../index/w/wildflower-honey-semifreddo-with-honey-sesame-wafers-232643.json)
* [Wilted Autumn Greens With Honey Mustard Vinaigrette](../../index/w/wilted-autumn-greens-with-honey-mustard-vinaigrette-51211620.json)
* [Yogurt With Fresh Figs Honey And Pine Nuts](../../index/y/yogurt-with-fresh-figs-honey-and-pine-nuts-51239840.json)
* [Honey Braised Beef Brisket](../../index/h/honey-braised-beef-brisket.json)
* [Honey Tangerine Fizz](../../index/h/honey-tangerine-fizz.json)
* [Sous Vide Beef With Honey Spices](../../index/s/sous-vide-beef-with-honey-spices.json)
* [Baked Ham With Brown Sugar Honey Glaze](../../index/b/baked-ham-with-brown-sugar-honey-glaze.json)
* [Baked Salmon With Honey Mustard Sauce](../../index/b/baked-salmon-with-honey-mustard-sauce.json)
* [Buttermilk Biscuits With Butter And Honey Recipe](../../index/b/buttermilk-biscuits-with-butter-and-honey-recipe.json)
* [Grilled Honey Teriyaki Chicken Recipe](../../index/g/grilled-honey-teriyaki-chicken-recipe.json)
* [Honey Cornbread Muffins Recipe](../../index/h/honey-cornbread-muffins-recipe.json)
* [Honey Glazed Carrots Recipe](../../index/h/honey-glazed-carrots-recipe.json)
* [Honey Glazed Ham](../../index/h/honey-glazed-ham.json)
* [Honey Ricotta Cheesecake Recipe](../../index/h/honey-ricotta-cheesecake-recipe.json)
* [Honey Soy Grilled Salmon With Edamame Recipe](../../index/h/honey-soy-grilled-salmon-with-edamame-recipe.json)
* [Orange Honey Glazed Carrots](../../index/o/orange-honey-glazed-carrots.json)
* [Pan Fried Honey Lemon Fingerling Potatoes And Leeks](../../index/p/pan-fried-honey-lemon-fingerling-potatoes-and-leeks.json)
* [Panna Cotta With Honey Recipe](../../index/p/panna-cotta-with-honey-recipe.json)
* [Roasted Sweet Potatoes With Honey Butter Recipe](../../index/r/roasted-sweet-potatoes-with-honey-butter-recipe.json)
* [Sticky Honey Soy Chicken Wings Recipe](../../index/s/sticky-honey-soy-chicken-wings-recipe.json)
* [Ancho Honey Mustard Sauce](../../index/a/ancho-honey-mustard-sauce.json)
* [Ancient Honey Cakes Rice Flour Cooki](../../index/a/ancient-honey-cakes-rice-flour-cooki.json)
* [Apple And Honey Sorbet](../../index/a/apple-and-honey-sorbet.json)
* [Apple Butter With Honey](../../index/a/apple-butter-with-honey.json)
* [Apple Honey Bundt Cake](../../index/a/apple-honey-bundt-cake.json)
* [Apple Honey Glaze](../../index/a/apple-honey-glaze.json)
* [Apple Honey Glazed Chicken](../../index/a/apple-honey-glazed-chicken.json)
* [Apricot And Honey Ham Glaze](../../index/a/apricot-and-honey-ham-glaze.json)
* [Ashley And Whitneys Honey Bbq Wings](../../index/a/ashley-and-whitneys-honey-bbq-wings.json)
* [Asian Inspired Honey Vanilla Chicken](../../index/a/asian-inspired-honey-vanilla-chicken.json)
* [Awesome Honey Pecan Pork Chops](../../index/a/awesome-honey-pecan-pork-chops.json)
* [Baked Asian Style Honey Chicken](../../index/b/baked-asian-style-honey-chicken.json)
* [Baked Honey Mustard Chicken](../../index/b/baked-honey-mustard-chicken.json)
* [Baked Honey Nut Chicken](../../index/b/baked-honey-nut-chicken.json)
* [Banana Honey Yogurt Ice](../../index/b/banana-honey-yogurt-ice.json)
* [Bean And Honey Burrito Casserole](../../index/b/bean-and-honey-burrito-casserole.json)
* [Belize Honey Bbq Sauce](../../index/b/belize-honey-bbq-sauce.json)
* [Bjs Carolina Honey Barbecue Sauce](../../index/b/bjs-carolina-honey-barbecue-sauce.json)
* [Bold Honey Barbecue Burger](../../index/b/bold-honey-barbecue-burger.json)
* [Bourbon Apple Cider And Honey Glazed](../../index/b/bourbon-apple-cider-and-honey-glazed.json)
* [Bread Machine Honey Oat Wheat Bread](../../index/b/bread-machine-honey-oat-wheat-bread.json)
* [Bries Banana And Honey Dog Treats](../../index/b/bries-banana-and-honey-dog-treats.json)
* [Brown Sugar And Honey Limeade](../../index/b/brown-sugar-and-honey-limeade.json)
* [Bruces Honey Sesame Bread](../../index/b/bruces-honey-sesame-bread.json)
* [Brussels Sprouts Slaw With Honey Yogu](../../index/b/brussels-sprouts-slaw-with-honey-yogu.json)
* [Butter Honey Wheat Bread](../../index/b/butter-honey-wheat-bread.json)
* [Butterfly Honey Pork Fillets](../../index/b/butterfly-honey-pork-fillets.json)
* [Buttermilk Honey Bread](../../index/b/buttermilk-honey-bread.json)
* [Buttermilk Honey Wheat Bread](../../index/b/buttermilk-honey-wheat-bread.json)
* [Canadian Honey Apple Walnut Cobbler](../../index/c/canadian-honey-apple-walnut-cobbler.json)
* [Cauliflower And Arugula Honey Sesame](../../index/c/cauliflower-and-arugula-honey-sesame.json)
* [Chef Johns Honey Glazed Ham](../../index/c/chef-johns-honey-glazed-ham.json)
* [Chicken Honey Mustard Pie](../../index/c/chicken-honey-mustard-pie.json)
* [Chicken Honey Nut Stir Fry](../../index/c/chicken-honey-nut-stir-fry.json)
* [Chipotle Honey Roasted Peanuts](../../index/c/chipotle-honey-roasted-peanuts.json)
* [Chocolate Chip Honey Cookies](../../index/c/chocolate-chip-honey-cookies.json)
* [Chocolate Glazed Honey Cake](../../index/c/chocolate-glazed-honey-cake.json)
* [Cinnamon Honey Butter](../../index/c/cinnamon-honey-butter.json)
* [Citrus Honey Brined Smoked Turkey](../../index/c/citrus-honey-brined-smoked-turkey.json)
* [Citrus Honey Green Tea](../../index/c/citrus-honey-green-tea.json)
* [Cordon Bleu Rollups With Honey Mustar](../../index/c/cordon-bleu-rollups-with-honey-mustar.json)
* [Country Style Honey Mustard](../../index/c/country-style-honey-mustard.json)
* [Cranberry Sauce With Honey And Pears](../../index/c/cranberry-sauce-with-honey-and-pears.json)
* [Creamy Goat Cheese And Honey](../../index/c/creamy-goat-cheese-and-honey.json)
* [Crispy Honey Garlic Sriracha Wings](../../index/c/crispy-honey-garlic-sriracha-wings.json)
* [Crispy Honey Sriracha Chicken Wings](../../index/c/crispy-honey-sriracha-chicken-wings.json)
* [Crunchy Honey Wheat Bread](../../index/c/crunchy-honey-wheat-bread.json)
* [Crusted Salmon With Honey Mustard Sau](../../index/c/crusted-salmon-with-honey-mustard-sau.json)
* [Curried Honey Mustard Chicken](../../index/c/curried-honey-mustard-chicken.json)
* [Curried Honey Mustard Sauce](../../index/c/curried-honey-mustard-sauce.json)
* [Dark Honey Pecan Granola](../../index/d/dark-honey-pecan-granola.json)
* [Dennis Honey Mustard Salad Dressing](../../index/d/dennis-honey-mustard-salad-dressing.json)
* [Detroit Hot Honey Wings](../../index/d/detroit-hot-honey-wings.json)
* [Dill And Honey Pork Chops](../../index/d/dill-and-honey-pork-chops.json)
* [Drumsticks With Peach And Honey](../../index/d/drumsticks-with-peach-and-honey.json)
* [Dsfs Honey Roasted Carrots And Parsn](../../index/d/dsfs-honey-roasted-carrots-and-parsn.json)
* [Duck With Honey Soy And Ginger](../../index/d/duck-with-honey-soy-and-ginger.json)
* [Easy And Good Honey Mustard Salad Dre](../../index/e/easy-and-good-honey-mustard-salad-dre.json)
* [Easy Apple Crisp With Honey](../../index/e/easy-apple-crisp-with-honey.json)
* [Easy Honey Garlic Chicken](../../index/e/easy-honey-garlic-chicken.json)
* [Easy Honey Mustard Ham](../../index/e/easy-honey-mustard-ham.json)
* [Easy Honey Mustard Mozzarella Chicken](../../index/e/easy-honey-mustard-mozzarella-chicken.json)
* [English Honey Roasted Turkey](../../index/e/english-honey-roasted-turkey.json)
* [Fat Free Mini Prune Honey Muffins](../../index/f/fat-free-mini-prune-honey-muffins.json)
* [Fennel In Wine And Honey](../../index/f/fennel-in-wine-and-honey.json)
* [Fig And Honey Jam With Walnuts](../../index/f/fig-and-honey-jam-with-walnuts.json)
* [Fish Tacos With Honey Cumin Cilantro](../../index/f/fish-tacos-with-honey-cumin-cilantro.json)
* [Flourless Banana Honey Cake](../../index/f/flourless-banana-honey-cake.json)
* [Freezer Pear Honey Jam](../../index/f/freezer-pear-honey-jam.json)
* [Fresh Apricot Honey Sponge Cake Russ](../../index/f/fresh-apricot-honey-sponge-cake-russ.json)
* [Fresh Fruit Salad With Honey Lime Dre](../../index/f/fresh-fruit-salad-with-honey-lime-dre.json)
* [Fresh Peaches With Honey Vanilla Crem](../../index/f/fresh-peaches-with-honey-vanilla-crem.json)
* [Fruit And Honey Cookies](../../index/f/fruit-and-honey-cookies.json)
* [Fruit N Honey Granola](../../index/f/fruit-n-honey-granola.json)
* [Ginger Honey Drunk Chili](../../index/g/ginger-honey-drunk-chili.json)
* [Glazed Turkey With Maille Honey Di](../../index/g/glazed-turkey-with-maille-honey-di.json)
* [Gluten Free Honey Scones](../../index/g/gluten-free-honey-scones.json)
* [Golden Honey Snaps](../../index/g/golden-honey-snaps.json)
* [Graces Honey Cakes](../../index/g/graces-honey-cakes.json)
* [Grandma Dalleys Honey Granola](../../index/g/grandma-dalleys-honey-granola.json)
* [Grandma Sadies Honey Cake](../../index/g/grandma-sadies-honey-cake.json)
* [Grandmas Honey Bun Cake](../../index/g/grandmas-honey-bun-cake.json)
* [Granola Honey](../../index/g/granola-honey.json)
* [Grape Honey Jam](../../index/g/grape-honey-jam.json)
* [Greek Honey Cake](../../index/g/greek-honey-cake.json)
* [Grilled Asian Honey Garlic Pork Tende](../../index/g/grilled-asian-honey-garlic-pork-tende.json)
* [Grilled Honey Chili Flap Steak](../../index/g/grilled-honey-chili-flap-steak.json)
* [Grilled Honey Lemon Chicken](../../index/g/grilled-honey-lemon-chicken.json)
* [Grilled Lime Cilantro Ahi With Honey](../../index/g/grilled-lime-cilantro-ahi-with-honey.json)
* [Ham Bagels With Honey Mustard Cream C](../../index/h/ham-bagels-with-honey-mustard-cream-c.json)
* [Ham With Honey And Brown Sugar Glaze](../../index/h/ham-with-honey-and-brown-sugar-glaze.json)
* [Healthy Honey Carrot Cookies](../../index/h/healthy-honey-carrot-cookies.json)
* [Herb N Honey Cucumber Tomato Salad](../../index/h/herb-n-honey-cucumber-tomato-salad.json)
* [Hoisin Honey Chicken Wings From Reyno](../../index/h/hoisin-honey-chicken-wings-from-reyno.json)
* [Holiday Honey Ham Glaze](../../index/h/holiday-honey-ham-glaze.json)
* [Holiday Turkey With Honey Orange Glaz](../../index/h/holiday-turkey-with-honey-orange-glaz.json)
* [Homemade Honey Peanut Butter](../../index/h/homemade-honey-peanut-butter.json)
* [Honey Almond Asparagus With Feta Chee](../../index/h/honey-almond-asparagus-with-feta-chee.json)
* [Honey Almond Chicken](../../index/h/honey-almond-chicken.json)
* [Honey Almond Seeded Granola](../../index/h/honey-almond-seeded-granola.json)
* [Honey And Citrus Dressing](../../index/h/honey-and-citrus-dressing.json)
* [Honey And Feta Spread](../../index/h/honey-and-feta-spread.json)
* [Honey And Flaxseed Bread](../../index/h/honey-and-flaxseed-bread.json)
* [Honey And Ginger Chicken](../../index/h/honey-and-ginger-chicken.json)
* [Honey And Nut Ice Cream](../../index/h/honey-and-nut-ice-cream.json)
* [Honey And Onion Mustard Sauce](../../index/h/honey-and-onion-mustard-sauce.json)
* [Honey And Rosemary Soda Bread](../../index/h/honey-and-rosemary-soda-bread.json)
* [Honey And Rosemary Sweet Potatoes](../../index/h/honey-and-rosemary-sweet-potatoes.json)
* [Honey Angel Food Cake](../../index/h/honey-angel-food-cake.json)
* [Honey Apple Brined Turkey](../../index/h/honey-apple-brined-turkey.json)
* [Honey Apricot Pork Chops](../../index/h/honey-apricot-pork-chops.json)
* [Honey Avocado Chicken Salad](../../index/h/honey-avocado-chicken-salad.json)
* [Honey Baked Apples](../../index/h/honey-baked-apples.json)
* [Honey Baked Chicken I](../../index/h/honey-baked-chicken-i.json)
* [Honey Baked Chicken Ii](../../index/h/honey-baked-chicken-ii.json)
* [Honey Baked Chicken](../../index/h/honey-baked-chicken.json)
* [Honey Baked Lentils](../../index/h/honey-baked-lentils.json)
* [Honey Balsamic Chicken](../../index/h/honey-balsamic-chicken.json)
* [Honey Balsamic Vinaigrette](../../index/h/honey-balsamic-vinaigrette.json)
* [Honey Barbecue Sauce](../../index/h/honey-barbecue-sauce.json)
* [Honey Bars](../../index/h/honey-bars.json)
* [Honey Bbq Sauce](../../index/h/honey-bbq-sauce.json)
* [Honey Bean Salad](../../index/h/honey-bean-salad.json)
* [Honey Bear Smoothie](../../index/h/honey-bear-smoothie.json)
* [Honey Beet Salad](../../index/h/honey-beet-salad.json)
* [Honey Black Pepper Chicken Wings](../../index/h/honey-black-pepper-chicken-wings.json)
* [Honey Bourbon Fire Sauce](../../index/h/honey-bourbon-fire-sauce.json)
* [Honey Bourbon Horseradish Glazed Ham](../../index/h/honey-bourbon-horseradish-glazed-ham.json)
* [Honey Bran Muffins](../../index/h/honey-bran-muffins.json)
* [Honey Brie Spread](../../index/h/honey-brie-spread.json)
* [Honey Brined Fried Chicken Breasts](../../index/h/honey-brined-fried-chicken-breasts.json)
* [Honey Brined Smoked Turkey](../../index/h/honey-brined-smoked-turkey.json)
* [Honey Brown Rolls Or Loaves](../../index/h/honey-brown-rolls-or-loaves.json)
* [Honey Bun Cake From Scratch](../../index/h/honey-bun-cake-from-scratch.json)
* [Honey Bun Cake I](../../index/h/honey-bun-cake-i.json)
* [Honey Bun Cake Ii](../../index/h/honey-bun-cake-ii.json)
* [Honey Bunch Bread](../../index/h/honey-bunch-bread.json)
* [Honey Butter Biscuit Cookies](../../index/h/honey-butter-biscuit-cookies.json)
* [Honey Butter Tarts](../../index/h/honey-butter-tarts.json)
* [Honey Butter Yule Ham](../../index/h/honey-butter-yule-ham.json)
* [Honey Butter Zucchini Bread](../../index/h/honey-butter-zucchini-bread.json)
* [Honey Butter](../../index/h/honey-butter.json)
* [Honey Buttermilk Oatbread](../../index/h/honey-buttermilk-oatbread.json)
* [Honey Cake I](../../index/h/honey-cake-i.json)
* [Honey Cake Ii](../../index/h/honey-cake-ii.json)
* [Honey Cake Iii](../../index/h/honey-cake-iii.json)
* [Honey Cake With Orange Glaze](../../index/h/honey-cake-with-orange-glaze.json)
* [Honey Carrot Cake](../../index/h/honey-carrot-cake.json)
* [Honey Carrots With A Hint Of Ginger](../../index/h/honey-carrots-with-a-hint-of-ginger.json)
* [Honey Chicken](../../index/h/honey-chicken.json)
* [Honey Chile Chicken](../../index/h/honey-chile-chicken.json)
* [Honey Chipotle Wing Sauce Glaze](../../index/h/honey-chipotle-wing-sauce-glaze.json)
* [Honey Chocolate Pie](../../index/h/honey-chocolate-pie.json)
* [Honey Citrus Turkey Brine](../../index/h/honey-citrus-turkey-brine.json)
* [Honey Clove Ham Glaze](../../index/h/honey-clove-ham-glaze.json)
* [Honey Coconut Salmon](../../index/h/honey-coconut-salmon.json)
* [Honey Comb Coffee Cake](../../index/h/honey-comb-coffee-cake.json)
* [Honey Cookies](../../index/h/honey-cookies.json)
* [Honey Corn Flake Amaranth Cookie](../../index/h/honey-corn-flake-amaranth-cookie.json)
* [Honey Cornbread Muffins](../../index/h/honey-cornbread-muffins.json)
* [Honey Cornbread](../../index/h/honey-cornbread.json)
* [Honey Cranberry Oat Bread](../../index/h/honey-cranberry-oat-bread.json)
* [Honey Crunch Pecan Pie](../../index/h/honey-crunch-pecan-pie.json)
* [Honey Crystal Almonds](../../index/h/honey-crystal-almonds.json)
* [Honey Curried Roasted Chicken And Veg](../../index/h/honey-curried-roasted-chicken-and-veg.json)
* [Honey Dijon Balsamic Vinaigrette](../../index/h/honey-dijon-balsamic-vinaigrette.json)
* [Honey Dijon Brussels Sprouts](../../index/h/honey-dijon-brussels-sprouts.json)
* [Honey Dijon Chicken With A Kick](../../index/h/honey-dijon-chicken-with-a-kick.json)
* [Honey Dijon Mustard And Poppy Seed Co](../../index/h/honey-dijon-mustard-and-poppy-seed-co.json)
* [Honey Dijon Pork Chops](../../index/h/honey-dijon-pork-chops.json)
* [Honey Dijon Vinaigrette](../../index/h/honey-dijon-vinaigrette.json)
* [Honey Dill Sauce](../../index/h/honey-dill-sauce.json)
* [Honey Dip](../../index/h/honey-dip.json)
* [Honey Dressing With Poppy Seeds](../../index/h/honey-dressing-with-poppy-seeds.json)
* [Honey Drizzled Curry Mustard Chicken](../../index/h/honey-drizzled-curry-mustard-chicken.json)
* [Honey Drops](../../index/h/honey-drops.json)
* [Honey Duck](../../index/h/honey-duck.json)
* [Honey Flaked Chicken](../../index/h/honey-flaked-chicken.json)
* [Honey Fried Chicken](../../index/h/honey-fried-chicken.json)
* [Honey Fruit Dessert](../../index/h/honey-fruit-dessert.json)
* [Honey Garlic Bbq Sauce Ii](../../index/h/honey-garlic-bbq-sauce-ii.json)
* [Honey Garlic Bbq Sauce](../../index/h/honey-garlic-bbq-sauce.json)
* [Honey Garlic Carrots](../../index/h/honey-garlic-carrots.json)
* [Honey Garlic Chicken Wings](../../index/h/honey-garlic-chicken-wings.json)
* [Honey Garlic Chicken With Rosemary](../../index/h/honey-garlic-chicken-with-rosemary.json)
* [Honey Garlic Chicken Wraps](../../index/h/honey-garlic-chicken-wraps.json)
* [Honey Garlic Glazed Salmon](../../index/h/honey-garlic-glazed-salmon.json)
* [Honey Garlic Moose Meatballs](../../index/h/honey-garlic-moose-meatballs.json)
* [Honey Garlic Ribs](../../index/h/honey-garlic-ribs.json)
* [Honey Garlic Sauce](../../index/h/honey-garlic-sauce.json)
* [Honey Garlic Slow Cooker Chicken Thig](../../index/h/honey-garlic-slow-cooker-chicken-thig.json)
* [Honey Garlic Vinaigrette](../../index/h/honey-garlic-vinaigrette.json)
* [Honey Ginger Carrots](../../index/h/honey-ginger-carrots.json)
* [Honey Ginger Green Beans](../../index/h/honey-ginger-green-beans.json)
* [Honey Ginger Grilled Salmon](../../index/h/honey-ginger-grilled-salmon.json)
* [Honey Ginger Kale Salad](../../index/h/honey-ginger-kale-salad.json)
* [Honey Ginger Orange Oatmeal Delight C](../../index/h/honey-ginger-orange-oatmeal-delight-c.json)
* [Honey Ginger Plum Sundae](../../index/h/honey-ginger-plum-sundae.json)
* [Honey Ginger Shrimp And Vegetables](../../index/h/honey-ginger-shrimp-and-vegetables.json)
* [Honey Ginger Shrimp](../../index/h/honey-ginger-shrimp.json)
* [Honey Glazed Brussels Sprouts](../../index/h/honey-glazed-brussels-sprouts.json)
* [Honey Glazed Carrots And Pears](../../index/h/honey-glazed-carrots-and-pears.json)
* [Honey Glazed Carrots](../../index/h/honey-glazed-carrots.json)
* [Honey Glazed Chicken](../../index/h/honey-glazed-chicken.json)
* [Honey Glazed Chinese Chicken](../../index/h/honey-glazed-chinese-chicken.json)
* [Honey Glazed Ham](../../index/h/honey-glazed-ham.json)
* [Honey Glazed Pea Pods And Carrots](../../index/h/honey-glazed-pea-pods-and-carrots.json)
* [Honey Glazed Pork Tenderloin](../../index/h/honey-glazed-pork-tenderloin.json)
* [Honey Glazed Salmon](../../index/h/honey-glazed-salmon.json)
* [Honey Glazed Scallops](../../index/h/honey-glazed-scallops.json)
* [Honey Glazed Stuffed Chicken](../../index/h/honey-glazed-stuffed-chicken.json)
* [Honey Graham Crackers](../../index/h/honey-graham-crackers.json)
* [Honey Grilled Chicken](../../index/h/honey-grilled-chicken.json)
* [Honey Grilled Pork Chops](../../index/h/honey-grilled-pork-chops.json)
* [Honey Grilled Shrimp](../../index/h/honey-grilled-shrimp.json)
* [Honey Ham Glaze](../../index/h/honey-ham-glaze.json)
* [Honey Ham](../../index/h/honey-ham.json)
* [Honey Hoisin Pork Chops](../../index/h/honey-hoisin-pork-chops.json)
* [Honey Jalapeno Burgers](../../index/h/honey-jalapeno-burgers.json)
* [Honey Key Lime Grilled Chicken](../../index/h/honey-key-lime-grilled-chicken.json)
* [Honey Kissed Pretzel Crusted Pork Loi](../../index/h/honey-kissed-pretzel-crusted-pork-loi.json)
* [Honey Lavender Cookies](../../index/h/honey-lavender-cookies.json)
* [Honey Lemon Ginger Tea](../../index/h/honey-lemon-ginger-tea.json)
* [Honey Lemon Glazed Salmon With Spinac](../../index/h/honey-lemon-glazed-salmon-with-spinac.json)
* [Honey Lemon Tea](../../index/h/honey-lemon-tea.json)
* [Honey Lime Chicken Enchiladas](../../index/h/honey-lime-chicken-enchiladas.json)
* [Honey Lime Chicken Wings](../../index/h/honey-lime-chicken-wings.json)
* [Honey Lime Enchiladas](../../index/h/honey-lime-enchiladas.json)
* [Honey Lime Fruit Salad](../../index/h/honey-lime-fruit-salad.json)
* [Honey Lime Fruit Toss](../../index/h/honey-lime-fruit-toss.json)
* [Honey Lime Granola With Almonds](../../index/h/honey-lime-granola-with-almonds.json)
* [Honey Lime Vinaigrette](../../index/h/honey-lime-vinaigrette.json)
* [Honey Limeade](../../index/h/honey-limeade.json)
* [Honey Macadamia Oat Bread](../../index/h/honey-macadamia-oat-bread.json)
* [Honey Milk Tea Hong Kong Style](../../index/h/honey-milk-tea---hong-kong-style.json)
* [Honey Multigrain Bread](../../index/h/honey-multigrain-bread.json)
* [Honey Mustard And Yogurt Salad Dres](../../index/h/honey-mustard-and-yogurt-salad-dres.json)
* [Honey Mustard Baked Chicken](../../index/h/honey-mustard-baked-chicken.json)
* [Honey Mustard Bbq Pork Chops](../../index/h/honey-mustard-bbq-pork-chops.json)
* [Honey Mustard Bbq Pork Medallions](../../index/h/honey-mustard-bbq-pork-medallions.json)
* [Honey Mustard Beer Brats](../../index/h/honey-mustard-beer-brats.json)
* [Honey Mustard Blend](../../index/h/honey-mustard-blend.json)
* [Honey Mustard Chicken With Pretzel Cr](../../index/h/honey-mustard-chicken-with-pretzel-cr.json)
* [Honey Mustard Chicken With Roasted Ve](../../index/h/honey-mustard-chicken-with-roasted-ve.json)
* [Honey Mustard Chicken](../../index/h/honey-mustard-chicken.json)
* [Honey Mustard Corned Beef Brisket And](../../index/h/honey-mustard-corned-beef-brisket-and.json)
* [Honey Mustard Crispy Chicken Wrap](../../index/h/honey-mustard-crispy-chicken-wrap.json)
* [Honey Mustard Dipping Sauce](../../index/h/honey-mustard-dipping-sauce.json)
* [Honey Mustard Dream Delight Pork Chop](../../index/h/honey-mustard-dream-delight-pork-chop.json)
* [Honey Mustard Dressing I](../../index/h/honey-mustard-dressing-i.json)
* [Honey Mustard Dressing Ii](../../index/h/honey-mustard-dressing-ii.json)
* [Honey Mustard Drizzle](../../index/h/honey-mustard-drizzle.json)
* [Honey Mustard Easy Franks](../../index/h/honey-mustard-easy-franks.json)
* [Honey Mustard Grilled Chicken](../../index/h/honey-mustard-grilled-chicken.json)
* [Honey Mustard Honey Mustard Dressing](../../index/h/honey-mustard-honey-mustard-dressing.json)
* [Honey Mustard Honey Mustard Dressing](../../index/h/honey-mustard-honey-mustard-dressing.json)
* [Honey Mustard Macaroni Salad](../../index/h/honey-mustard-macaroni-salad.json)
* [Honey Mustard Oven Baked Chicken Tend](../../index/h/honey-mustard-oven-baked-chicken-tend.json)
* [Honey Mustard Pork Chops](../../index/h/honey-mustard-pork-chops.json)
* [Honey Mustard Roasted Potatoes](../../index/h/honey-mustard-roasted-potatoes.json)
* [Honey Mustard Sauce For Dipping](../../index/h/honey-mustard-sauce-for-dipping.json)
* [Honey Mustard Sauce](../../index/h/honey-mustard-sauce.json)
* [Honey Mustard Soy Salmon](../../index/h/honey-mustard-soy-salmon.json)
* [Honey Mustard Steak](../../index/h/honey-mustard-steak.json)
* [Honey Mustard Stuffed Chicken Breasts](../../index/h/honey-mustard-stuffed-chicken-breasts.json)
* [Honey Mustard Tofu](../../index/h/honey-mustard-tofu.json)
* [Honey Nut Bars](../../index/h/honey-nut-bars.json)
* [Honey Nut Dipped Chicken](../../index/h/honey-nut-dipped-chicken.json)
* [Honey Nut Fruit Muffins](../../index/h/honey-nut-fruit-muffins.json)
* [Honey Nut Granola](../../index/h/honey-nut-granola.json)
* [Honey Nutters](../../index/h/honey-nutters.json)
* [Honey Nutty Granola](../../index/h/honey-nutty-granola.json)
* [Honey Oat Beer Bread](../../index/h/honey-oat-beer-bread.json)
* [Honey Oatmeal Bread I](../../index/h/honey-oatmeal-bread-i.json)
* [Honey Oatmeal Bread Ii](../../index/h/honey-oatmeal-bread-ii.json)
* [Honey Oatmeal Cookies](../../index/h/honey-oatmeal-cookies.json)
* [Honey Of An Oatmeal Bread](../../index/h/honey-of-an-oatmeal-bread.json)
* [Honey Orange Basil Chicken](../../index/h/honey-orange-basil-chicken.json)
* [Honey Orange Bison Back Ribs](../../index/h/honey-orange-bison-back-ribs.json)
* [Honey Orange Cornish Hens](../../index/h/honey-orange-cornish-hens.json)
* [Honey Orange Firecracker Shrimp](../../index/h/honey-orange-firecracker-shrimp.json)
* [Honey Orange Glazed Ham](../../index/h/honey-orange-glazed-ham.json)
* [Honey Orange Green Beans](../../index/h/honey-orange-green-beans.json)
* [Honey Orange Ham Glaze](../../index/h/honey-orange-ham-glaze.json)
* [Honey Orange Marmalade](../../index/h/honey-orange-marmalade.json)
* [Honey Pancake Puff](../../index/h/honey-pancake-puff.json)
* [Honey Parsnips](../../index/h/honey-parsnips.json)
* [Honey Peach Pie](../../index/h/honey-peach-pie.json)
* [Honey Peanut Granola](../../index/h/honey-peanut-granola.json)
* [Honey Peanut School Fuel](../../index/h/honey-peanut-school-fuel.json)
* [Honey Pecan Pumpkin Pie](../../index/h/honey-pecan-pumpkin-pie.json)
* [Honey Pie From Sifnos](../../index/h/honey-pie-from-sifnos.json)
* [Honey Pineapple Chicken](../../index/h/honey-pineapple-chicken.json)
* [Honey Poppy Seed Cornish Hens](../../index/h/honey-poppy-seed-cornish-hens.json)
* [Honey Poppy Seed Dressing](../../index/h/honey-poppy-seed-dressing.json)
* [Honey Pork Fillets](../../index/h/honey-pork-fillets.json)
* [Honey Pork Oriental](../../index/h/honey-pork-oriental.json)
* [Honey Pork With Peppers](../../index/h/honey-pork-with-peppers.json)
* [Honey Pumpkin Seeds](../../index/h/honey-pumpkin-seeds.json)
* [Honey Red Pepper Glazed Back Ribs](../../index/h/honey-red-pepper-glazed-back-ribs.json)
* [Honey Ribs](../../index/h/honey-ribs.json)
* [Honey Rice](../../index/h/honey-rice.json)
* [Honey Roasted Almonds](../../index/h/honey-roasted-almonds.json)
* [Honey Roasted Carrots With Cumin](../../index/h/honey-roasted-carrots-with-cumin.json)
* [Honey Roasted Carrots](../../index/h/honey-roasted-carrots.json)
* [Honey Roasted Peanuts](../../index/h/honey-roasted-peanuts.json)
* [Honey Roasted Potatoes And Mushrooms](../../index/h/honey-roasted-potatoes-and-mushrooms.json)
* [Honey Roasted Red Potatoes](../../index/h/honey-roasted-red-potatoes.json)
* [Honey Roasted Sweet Potatoes](../../index/h/honey-roasted-sweet-potatoes.json)
* [Honey Rosemary Carrots](../../index/h/honey-rosemary-carrots.json)
* [Honey Skirt Steak Appetizer With Hors](../../index/h/honey-skirt-steak-appetizer-with-hors.json)
* [Honey Smoked Turkey](../../index/h/honey-smoked-turkey.json)
* [Honey Sourdough](../../index/h/honey-sourdough.json)
* [Honey Soy Pork Chops From The Bradsha](../../index/h/honey-soy-pork-chops-from-the-bradsha.json)
* [Honey Soy Sweet Potatoes](../../index/h/honey-soy-sweet-potatoes.json)
* [Honey Soy Tilapia](../../index/h/honey-soy-tilapia.json)
* [Honey Spice Drumsticks](../../index/h/honey-spice-drumsticks.json)
* [Honey Spice Whole Wheat Banana Bread](../../index/h/honey-spice-whole-wheat-banana-bread.json)
* [Honey Spiced Granola Bars](../../index/h/honey-spiced-granola-bars.json)
* [Honey Spiced Peaches](../../index/h/honey-spiced-peaches.json)
* [Honey Steelhead Trout](../../index/h/honey-steelhead-trout.json)
* [Honey Sweetened Granola](../../index/h/honey-sweetened-granola.json)
* [Honey Syrup Date Cake](../../index/h/honey-syrup-date-cake.json)
* [Honey Syrup](../../index/h/honey-syrup.json)
* [Honey Vanilla Cheesecake With Gingers](../../index/h/honey-vanilla-cheesecake-with-gingers.json)
* [Honey Vanilla Grilled Sweet Potatoes](../../index/h/honey-vanilla-grilled-sweet-potatoes.json)
* [Honey Vanilla Ice Cream](../../index/h/honey-vanilla-ice-cream.json)
* [Honey Vanilla Pudding](../../index/h/honey-vanilla-pudding.json)
* [Honey Walnut Cream Cheese](../../index/h/honey-walnut-cream-cheese.json)
* [Honey Walnut Shrimp](../../index/h/honey-walnut-shrimp.json)
* [Honey Wheat Bread I](../../index/h/honey-wheat-bread-i.json)
* [Honey Wheat Bread Ii](../../index/h/honey-wheat-bread-ii.json)
* [Honey Wheat Bread Iii](../../index/h/honey-wheat-bread-iii.json)
* [Honey Wheat Bread Iv](../../index/h/honey-wheat-bread-iv.json)
* [Honey Wheat Cookies](../../index/h/honey-wheat-cookies.json)
* [Honey Wheat Muffins](../../index/h/honey-wheat-muffins.json)
* [Honey Wheat Oat Flour Bread Machine B](../../index/h/honey-wheat-oat-flour-bread-machine-b.json)
* [Honey Wheat Sandwich Rolls](../../index/h/honey-wheat-sandwich-rolls.json)
* [Honey White Bread](../../index/h/honey-white-bread.json)
* [Honey Whole Wheat Bread](../../index/h/honey-whole-wheat-bread.json)
* [Honey Whole Wheat Challah](../../index/h/honey-whole-wheat-challah.json)
* [Horseradish And Honey Glazed Ham](../../index/h/horseradish-and-honey-glazed-ham.json)
* [Hot Honey Whiskey Bbq Wing Sauce](../../index/h/hot-honey-whiskey-bbq-wing-sauce.json)
* [Hot Salmon Salad With Maille Honey](../../index/h/hot-salmon-salad-with-maille-honey.json)
* [Justins Honey Buffalo Chicken Wraps](../../index/j/justins-honey-buffalo-chicken-wraps.json)
* [Kale Chips With Honey](../../index/k/kale-chips-with-honey.json)
* [Kevins Toasted Honey Wheat Berry Bol](../../index/k/kevins-toasted-honey-wheat-berry-bol.json)
* [Kielbasa With Honey Mustard](../../index/k/kielbasa-with-honey-mustard.json)
* [Kristas Sticky Honey Garlic Wings](../../index/k/kristas-sticky-honey-garlic-wings.json)
* [Lamb Ribs With Honey And Wine](../../index/l/lamb-ribs-with-honey-and-wine.json)
* [Lamb Shanks With Ancho Chile Honey Gl](../../index/l/lamb-shanks-with-ancho-chile-honey-gl.json)
* [Latvian Honey Cake](../../index/l/latvian-honey-cake.json)
* [Lemon Honey Butter Tarts](../../index/l/lemon-honey-butter-tarts.json)
* [Lemon Honey Glazed Carrots](../../index/l/lemon-honey-glazed-carrots.json)
* [Lemon Honey Lamb Skewers](../../index/l/lemon-honey-lamb-skewers.json)
* [Lemon Honey Muffins](../../index/l/lemon-honey-muffins.json)
* [Lime Honey Soda](../../index/l/lime-honey-soda.json)
* [Maille Honey Dijon Chicken](../../index/m/maille-honey-dijon-chicken.json)
* [Margarets Honey Mustard](../../index/m/margarets-honey-mustard.json)
* [Medenjaci Croatian Honey Spice Cooki](../../index/m/medenjaci-croatian-honey-spice-cooki.json)
* [Melt In Your Mouth Honey Lemon Merry](../../index/m/melt-in-your-mouth-honey-lemon-merry.json)
* [Meyer Lemon Olive Oil And Honey Chevr](../../index/m/meyer-lemon-olive-oil-and-honey-chevr.json)
* [Milk And Honey Loaf](../../index/m/milk-and-honey-loaf.json)
* [Miso Honey Dressing](../../index/m/miso-honey-dressing.json)
* [Mixed Berry Pie With Honey Whole Whea](../../index/m/mixed-berry-pie-with-honey-whole-whea.json)
* [Moroccan Lamb With Shiraz Honey Sauce](../../index/m/moroccan-lamb-with-shiraz-honey-sauce.json)
* [Mrs Carrigans Honey Wheat Bread](../../index/m/mrs-carrigans-honey-wheat-bread.json)
* [Mushroom Honey Lager Sauce](../../index/m/mushroom-honey-lager-sauce.json)
* [No Bake Energy Bars With Honey](../../index/n/no-bake-energy-bars-with-honey.json)
* [Non Fat Honey Mustard Dressing](../../index/n/non-fat-honey-mustard-dressing.json)
* [Oat N Honey Bread](../../index/o/oat-n-honey-bread.json)
* [Oatmeal Rolls With Honey Butter](../../index/o/oatmeal-rolls-with-honey-butter.json)
* [Old Fashioned Honey Pecan Pie](../../index/o/old-fashioned-honey-pecan-pie.json)
* [Old German Honey Cookies](../../index/o/old-german-honey-cookies.json)
* [Orange Honey And Soy Chicken](../../index/o/orange-honey-and-soy-chicken.json)
* [Orange Honey Butter](../../index/o/orange-honey-butter.json)
* [Orange Honey Garlic Chicken](../../index/o/orange-honey-garlic-chicken.json)
* [Orange Juice Muffins With Honey Sprea](../../index/o/orange-juice-muffins-with-honey-sprea.json)
* [Orange Soy And Honey Pulled Pork](../../index/o/orange-soy-and-honey-pulled-pork.json)
* [Passover Apples And Honey Charoset](../../index/p/passover-apples-and-honey-charoset.json)
* [Peanut Butter And Honey Balls](../../index/p/peanut-butter-and-honey-balls.json)
* [Peanut Butter And Honey Granola Bars](../../index/p/peanut-butter-and-honey-granola-bars.json)
* [Peanut Butter And Honey No Bake Cooki](../../index/p/peanut-butter-and-honey-no-bake-cooki.json)
* [Peanut Butter And Honey Overnight Oat](../../index/p/peanut-butter-and-honey-overnight-oat.json)
* [Peanut Butter And Honey Sandwich](../../index/p/peanut-butter-and-honey-sandwich.json)
* [Peanut Butter Bacon And Honey Sandwi](../../index/p/peanut-butter-bacon-and-honey-sandwi.json)
* [Peanut Butter Honey Balls](../../index/p/peanut-butter-honey-balls.json)
* [Pear Honey Cranberry Sauce](../../index/p/pear-honey-cranberry-sauce.json)
* [Pear Honey](../../index/p/pear-honey.json)
* [Pecan Honey Glazed Fried Chicken](../../index/p/pecan-honey-glazed-fried-chicken.json)
* [Pepper Honey Cedar Plank Salmon](../../index/p/pepper-honey-cedar-plank-salmon.json)
* [Piernik Honey Bread](../../index/p/piernik---honey-bread.json)
* [Pina Colada Power Honey Smoothie](../../index/p/pina-colada-power-honey-smoothie.json)
* [Pork Chops With Tangy Honey Sauce](../../index/p/pork-chops-with-tangy-honey-sauce.json)
* [Pork Tenderloin With A Honey Grape Sa](../../index/p/pork-tenderloin-with-a-honey-grape-sa.json)
* [Pumpkin Honey Butter](../../index/p/pumpkin-honey-butter.json)
* [Pumpkin Honey Pie](../../index/p/pumpkin-honey-pie.json)
* [Pumpkin Wheat Honey Muffins](../../index/p/pumpkin-wheat-honey-muffins.json)
* [Quinoa And Honey Mustard Chicken Slow](../../index/q/quinoa-and-honey-mustard-chicken-slow.json)
* [Raisin Honey Chews](../../index/r/raisin-honey-chews.json)
* [Ricos Passionate Pink Honey Lemonade](../../index/r/ricos-passionate-pink-honey-lemonade.json)
* [Roasted Carrots With Honey Sriracha S](../../index/r/roasted-carrots-with-honey-sriracha-s.json)
* [Roasted Orange Rosemary Honey Glazed](../../index/r/roasted-orange-rosemary-honey-glazed.json)
* [Salmon With Pecan Honey Glaze](../../index/s/salmon-with-pecan-honey-glaze.json)
* [Skillet Honey Beef](../../index/s/skillet-honey-beef.json)
* [Slow Cooker Apple Butter With Honey](../../index/s/slow-cooker-apple-butter-with-honey.json)
* [Slow Cooker Honey Garlic Chicken](../../index/s/slow-cooker-honey-garlic-chicken.json)
* [Slow Cooker Parmesan Honey Pork Roast](../../index/s/slow-cooker-parmesan-honey-pork-roast.json)
* [Southern Style Honey Butter](../../index/s/southern-style-honey-butter.json)
* [Soy And Honey Pork Chops](../../index/s/soy-and-honey-pork-chops.json)
* [Soy Honey Glazed Salmon With Asparagu](../../index/s/soy-honey-glazed-salmon-with-asparagu.json)
* [Special Honey Bbq Sauce](../../index/s/special-honey-bbq-sauce.json)
* [Spicy Honey Asian Chops](../../index/s/spicy-honey-asian-chops.json)
* [Spicy Honey Mustard Glazed Pork Ribs](../../index/s/spicy-honey-mustard-glazed-pork-ribs.json)
* [Spicy Honey Mustard Pork Roast](../../index/s/spicy-honey-mustard-pork-roast.json)
* [Spicy Honey Mustard Sauce](../../index/s/spicy-honey-mustard-sauce.json)
* [Spicy Honey Roasted Chicken](../../index/s/spicy-honey-roasted-chicken.json)
* [Spinach And Fruit Honey Salad](../../index/s/spinach-and-fruit-honey-salad.json)
* [Sriracha Honey Chicken Legs](../../index/s/sriracha-honey-chicken-legs.json)
* [Sriracha Honey Chicken Pizza](../../index/s/sriracha-honey-chicken-pizza.json)
* [Sriracha Honey Chicken](../../index/s/sriracha-honey-chicken.json)
* [Sriracha Honey Salmon](../../index/s/sriracha-honey-salmon.json)
* [Strawberry Salad With Shallot Honey V](../../index/s/strawberry-salad-with-shallot-honey-v.json)
* [Striped Bass Seared With Honey And Ou](../../index/s/striped-bass-seared-with-honey-and-ou.json)
* [Succulent Grilled Peaches With Honey](../../index/s/succulent-grilled-peaches-with-honey.json)
* [Summer Asian Salad With Honey Chicken](../../index/s/summer-asian-salad-with-honey-chicken.json)
* [Super Crunch Oven Cooked Honey Dipped](../../index/s/super-crunch-oven-cooked-honey-dipped.json)
* [Sweet And Sour Honey Mustard Pork Ten](../../index/s/sweet-and-sour-honey-mustard-pork-ten.json)
* [Sweet Honey French Bread](../../index/s/sweet-honey-french-bread.json)
* [Tahini Honey Oatmeal Cookies](../../index/t/tahini-honey-oatmeal-cookies.json)
* [Tangy Honey Glazed Ham](../../index/t/tangy-honey-glazed-ham.json)
* [Thai Style Peanut Sauce With Honey](../../index/t/thai-style-peanut-sauce-with-honey.json)
* [The Best Homemade Honey Garlic Sauce](../../index/t/the-best-homemade-honey-garlic-sauce.json)
* [Turmeric Honey Paste](../../index/t/turmeric-honey-paste.json)
* [Twist Of Honey Orange Rolls](../../index/t/twist-of-honey-orange-rolls.json)
* [Vanilla Honey Simple Syrup](../../index/v/vanilla-honey-simple-syrup.json)
* [Vanilla Port Poached Figs With Honey](../../index/v/vanilla-port-poached-figs-with-honey.json)
* [Viryta Lithuanian Honey Liqueur](../../index/v/viryta-lithuanian-honey-liqueur.json)
* [Warm Jeweled Barley Salad With Honey ](../../index/w/warm-jeweled-barley-salad-with-honey-.json)
* [Warm Lemon Honey And Ginger Soother](../../index/w/warm-lemon-honey-and-ginger-soother.json)
* [Whipped Ricotta With Honey And Summer](../../index/w/whipped-ricotta-with-honey-and-summer.json)
* [Whole Wheat And Honey Pizza Dough](../../index/w/whole-wheat-and-honey-pizza-dough.json)
* [Whole Wheat Honey Bread](../../index/w/whole-wheat-honey-bread.json)
* [Whole White Wheat And Honey Chocolate](../../index/w/whole-white-wheat-and-honey-chocolate.json)
* [Worlds Best Honey Garlic Pork Chops](../../index/w/worlds-best-honey-garlic-pork-chops.json)
* [Yummy Honey Chicken Kabobs](../../index/y/yummy-honey-chicken-kabobs.json)
* [Yummy Honey Mustard Dipping Sauce](../../index/y/yummy-honey-mustard-dipping-sauce.json)
* [Apple Goat Cheese And Honey Tartlets 240112](../../index/a/apple-goat-cheese-and-honey-tartlets-240112.json)
* [Apple Streusel Tart With Honey Crust 233182](../../index/a/apple-streusel-tart-with-honey-crust-233182.json)
* [Apricot Berry And Jicama Salad With Honey Lime Dressing 106615](../../index/a/apricot-berry-and-jicama-salad-with-honey-lime-dressing-106615.json)
* [Apricot Tart With Honey And Almonds 101604](../../index/a/apricot-tart-with-honey-and-almonds-101604.json)
* [Apricot Yogurt Cake With Orange Honey Syrup 12086](../../index/a/apricot-yogurt-cake-with-orange-honey-syrup-12086.json)
* [Apricots Yogurt And Honey 239226](../../index/a/apricots-yogurt-and-honey-239226.json)
* [Baked Apples Stuffed With Honey Almonds And Ginger 232891](../../index/b/baked-apples-stuffed-with-honey-almonds-and-ginger-232891.json)
* [Baked Pears With Honey And Ginger 538](../../index/b/baked-pears-with-honey-and-ginger-538.json)
* [Balsamic Roasted Acorn Squash With Hot Chiles And Honey 108826](../../index/b/balsamic-roasted-acorn-squash-with-hot-chiles-and-honey-108826.json)
* [Bbq Onion Steaks With Honey Mustard Sauce 235354](../../index/b/bbq-onion-steaks-with-honey-mustard-sauce-235354.json)
* [Beef Short Ribs Tagine With Honey Glazed Butternut Squash 4563](../../index/b/beef-short-ribs-tagine-with-honey-glazed-butternut-squash-4563.json)
* [Beet Cucumber And Sweet Onion Salad With Dijon Honey Dressing 106820](../../index/b/beet-cucumber-and-sweet-onion-salad-with-dijon-honey-dressing-106820.json)
* [Belgian Endive Fuyu Persimmon And Baby Spinach Salad With Honey Mustard Vinaigrette 236129](../../index/b/belgian-endive-fuyu-persimmon-and-baby-spinach-salad-with-honey-mustard-vinaigrette-236129.json)
* [Blackberry Honey And Yogurt Pops 360253](../../index/b/blackberry-honey-and-yogurt-pops-360253.json)
* [Blue Cheese Apple And Sage Honey Crostini 51255310](../../index/b/blue-cheese-apple-and-sage-honey-crostini-51255310.json)
* [Blue Cheese With Rosemary Honey On Crackers 350621](../../index/b/blue-cheese-with-rosemary-honey-on-crackers-350621.json)
* [Bread Baked With Honey And Cream 106080](../../index/b/bread-baked-with-honey-and-cream-106080.json)
* [Brown Sugar Chipotle Salmon With Honey Berry Glaze 231282](../../index/b/brown-sugar-chipotle-salmon-with-honey-berry-glaze-231282.json)
* [Buttermilk And Honey Sorbet In Grapefruit Cups 2074](../../index/b/buttermilk-and-honey-sorbet-in-grapefruit-cups-2074.json)
* [Buttermilk Biscuits With Honey Butter](../../index/b/buttermilk-biscuits-with-honey-butter.json)
* [Butternut Squash Puree With Orange Ginger And Honey 107299](../../index/b/butternut-squash-puree-with-orange-ginger-and-honey-107299.json)
* [Caramelized Honey Nut And Seed Tart 56390106](../../index/c/caramelized-honey-nut-and-seed-tart-56390106.json)
* [Cardamom Yogurt Pudding With Orange And Cinnamon Honey Syrup 109474](../../index/c/cardamom-yogurt-pudding-with-orange-and-cinnamon-honey-syrup-109474.json)
* [Carrot Snow Pea And Red Pepper Julienne In Honey Vinaigrette 14006](../../index/c/carrot-snow-pea-and-red-pepper-julienne-in-honey-vinaigrette-14006.json)
* [Carrots And Rutabagas With Lemon And Honey 105812](../../index/c/carrots-and-rutabagas-with-lemon-and-honey-105812.json)
* [Cheddar Scones With Ham And Honey Mustard Butter 101500](../../index/c/cheddar-scones-with-ham-and-honey-mustard-butter-101500.json)
* [Chicken Liver Mousse With Burnt Honey Gelee](../../index/c/chicken-liver-mousse-with-burnt-honey-gelee.json)
* [Chicken Wings With Honey Soy Glaze 5619](../../index/c/chicken-wings-with-honey-soy-glaze-5619.json)
* [Chicken With Honey Mustard Sauce 5718](../../index/c/chicken-with-honey-mustard-sauce-5718.json)
* [Chicken With Tangerine Honey And Chipotle Glaze 232383](../../index/c/chicken-with-tangerine-honey-and-chipotle-glaze-232383.json)
* [Chili And Honey Chicken Legs 231997](../../index/c/chili-and-honey-chicken-legs-231997.json)
* [Chipotle Honey Glazed Chicken Wings With Toasted Sesame Seeds And Green Onion 363309](../../index/c/chipotle-honey-glazed-chicken-wings-with-toasted-sesame-seeds-and-green-onion-363309.json)
* [Chipotle Honey Glazed Pork 102275](../../index/c/chipotle-honey-glazed-pork-102275.json)
* [Chocolate Honey Dome Cake With Chocolate Honey Glaze 237916](../../index/c/chocolate-honey-dome-cake-with-chocolate-honey-glaze-237916.json)
* [Chocolate Honey Dome Cake With Chocolate Honey Glaze 237916](../../index/c/chocolate-honey-dome-cake-with-chocolate-honey-glaze-237916.json)
* [Chocolate Orange And Honey Cake 1630](../../index/c/chocolate-orange-and-honey-cake-1630.json)
* [Chopped Honey Mustard Slaw 232369](../../index/c/chopped-honey-mustard-slaw-232369.json)
* [Citrus Jell O With Honey And Mint 237833](../../index/c/citrus-jell-o-with-honey-and-mint-237833.json)
* [Coffee And Turron Ice Cream Torte With Honey Brandy Oranges 107168](../../index/c/coffee-and-turron-ice-cream-torte-with-honey-brandy-oranges-107168.json)
* [Cornish Game Hens With Dried Fruit And Honey 105525](../../index/c/cornish-game-hens-with-dried-fruit-and-honey-105525.json)
* [Cornmeal Pancakes With Honey Pecan Butter 108554](../../index/c/cornmeal-pancakes-with-honey-pecan-butter-108554.json)
* [Cream Tart With Oranges Honey And Toasted Almond Crust 357275](../../index/c/cream-tart-with-oranges-honey-and-toasted-almond-crust-357275.json)
* [Creme Fraiche Cheesecake With Honey Rum Roasted Pineapple 109089](../../index/c/creme-fraiche-cheesecake-with-honey-rum-roasted-pineapple-109089.json)
* [Crispy Lamb Spare Ribs With Honey Creme Fraiche And Chopped Pea Salad 241226](../../index/c/crispy-lamb-spare-ribs-with-honey-creme-fraiche-and-chopped-pea-salad-241226.json)
* [Dark Ginger Rye Cake With Yogurt And Honey](../../index/d/dark-ginger-rye-cake-with-yogurt-and-honey.json)
* [Date And Walnut Phyllo Rolls With Greek Yogurt And Honey 233722](../../index/d/date-and-walnut-phyllo-rolls-with-greek-yogurt-and-honey-233722.json)
* [Duck Breasts With Orange Honey And Tea Sauce 103268](../../index/d/duck-breasts-with-orange-honey-and-tea-sauce-103268.json)
* [Duck With Honey 101622](../../index/d/duck-with-honey-101622.json)
* [Duck With Honey Peppercorn Sauce 4690](../../index/d/duck-with-honey-peppercorn-sauce-4690.json)
* [Earl Grey Tea Madeleines With Honey 231609](../../index/e/earl-grey-tea-madeleines-with-honey-231609.json)
* [Feta With Pepper Honey 100908](../../index/f/feta-with-pepper-honey-100908.json)
* [Feta With Rosemary Pepper Honey 13096](../../index/f/feta-with-rosemary-pepper-honey-13096.json)
* [Figs And Dried Cherries Poached In Vanilla Honey Syrup With Pistachios 107510](../../index/f/figs-and-dried-cherries-poached-in-vanilla-honey-syrup-with-pistachios-107510.json)
* [Figs With Honey Orange Mascarpone And Pistachio Brittle 361232](../../index/f/figs-with-honey-orange-mascarpone-and-pistachio-brittle-361232.json)
* [Fluffy Cathead Biscuits With Honey Butter](../../index/f/fluffy-cathead-biscuits-with-honey-butter.json)
* [Fresh Fruit Salad With Honey Mint And Lime Syrup 15546](../../index/f/fresh-fruit-salad-with-honey-mint-and-lime-syrup-15546.json)
* [Fresh Fruit With Honey Yogurt Dip 5457](../../index/f/fresh-fruit-with-honey-yogurt-dip-5457.json)
* [Fried Chicken And Biscuits With Pink Peppercorn Honey Butter 354874](../../index/f/fried-chicken-and-biscuits-with-pink-peppercorn-honey-butter-354874.json)
* [Fried Eggplant With Honey Mint And Sesame Seeds 107127](../../index/f/fried-eggplant-with-honey-mint-and-sesame-seeds-107127.json)
* [Fried Pastry Spirals With Honey Sesame And Walnuts 108130](../../index/f/fried-pastry-spirals-with-honey-sesame-and-walnuts-108130.json)
* [Frozen Honey Mousse With Tart Cherry Cookies 364675](../../index/f/frozen-honey-mousse-with-tart-cherry-cookies-364675.json)
* [Fruit Salad With Honey Lime Dressing 106941](../../index/f/fruit-salad-with-honey-lime-dressing-106941.json)
* [Gala Apple Newtons With Honey Cream 108699](../../index/g/gala-apple-newtons-with-honey-cream-108699.json)
* [Game Hens With Honey Vinegar And Spice Glaze 1886](../../index/g/game-hens-with-honey-vinegar-and-spice-glaze-1886.json)
* [Ginger And Honey Baby Back Ribs 353813](../../index/g/ginger-and-honey-baby-back-ribs-353813.json)
* [Ginger Honey Lemonade 201200](../../index/g/ginger-honey-lemonade-201200.json)
* [Goat Cheese And Honey Blancmange 240232](../../index/g/goat-cheese-and-honey-blancmange-240232.json)
* [Goat Cheese Toasts With Walnuts Honey Thyme 51166200](../../index/g/goat-cheese-toasts-with-walnuts-honey-thyme-51166200.json)
* [Greek Honey And Anise Twists 3051](../../index/g/greek-honey-and-anise-twists-3051.json)
* [Grilled Corn With Honey Ginger Barbecue Sauce 51108010](../../index/g/grilled-corn-with-honey-ginger-barbecue-sauce-51108010.json)
* [Grilled Lamb Chops With Red Wine Garlic And Honey Glaze 5610](../../index/g/grilled-lamb-chops-with-red-wine-garlic-and-honey-glaze-5610.json)
* [Grilled Lemon Honey Chicken Breasts 101958](../../index/g/grilled-lemon-honey-chicken-breasts-101958.json)
* [Grilled Nectarines With Honey Balsamic Glaze 106651](../../index/g/grilled-nectarines-with-honey-balsamic-glaze-106651.json)
* [Grilled Pineapple And Banana Sauce With Honey And Lime 239044](../../index/g/grilled-pineapple-and-banana-sauce-with-honey-and-lime-239044.json)
* [Grilled Quail Salad With Honey Dijon Dressing 239843](../../index/g/grilled-quail-salad-with-honey-dijon-dressing-239843.json)
* [Grilled Shrimp With Honey Ginger Barbecue Sauce 51108030](../../index/g/grilled-shrimp-with-honey-ginger-barbecue-sauce-51108030.json)
* [Grilled Skewered Scallops And Apricots With Honey Mustard Dressing 235338](../../index/g/grilled-skewered-scallops-and-apricots-with-honey-mustard-dressing-235338.json)
* [Ham And Munster Buns With Honey Mustard 10545](../../index/h/ham-and-munster-buns-with-honey-mustard-10545.json)
* [Ham Stuffed Chicken Legs With Sweet Pepper Honey And Onion Compote 1964](../../index/h/ham-stuffed-chicken-legs-with-sweet-pepper-honey-and-onion-compote-1964.json)
* [High Honey Sriracha Popcorn](../../index/h/high-honey-sriracha-popcorn.json)
* [Hoisin And Honey Glazed Pork Chops 109015](../../index/h/hoisin-and-honey-glazed-pork-chops-109015.json)
* [Hoisin And Honey Pork Riblets 12141](../../index/h/hoisin-and-honey-pork-riblets-12141.json)
* [Honey Almond And Date Ice Cream Sauce 10391](../../index/h/honey-almond-and-date-ice-cream-sauce-10391.json)
* [Honey Almond Granola 350875](../../index/h/honey-almond-granola-350875.json)
* [Honey And Soy Glazed Salmon 14110](../../index/h/honey-and-soy-glazed-salmon-14110.json)
* [Honey And Spice Cranberry Sauce 100322](../../index/h/honey-and-spice-cranberry-sauce-100322.json)
* [Honey And Spice Loaf Cake 102698](../../index/h/honey-and-spice-loaf-cake-102698.json)
* [Honey And Thyme Ice Cream With Candied Thyme 236890](../../index/h/honey-and-thyme-ice-cream-with-candied-thyme-236890.json)
* [Honey And Vinegar Braised Short Ribs With Spinach 102858](../../index/h/honey-and-vinegar-braised-short-ribs-with-spinach-102858.json)
* [Honey Anise And Almond Biscotti 103158](../../index/h/honey-anise-and-almond-biscotti-103158.json)
* [Honey Barbequed Salmon With Mache 12667](../../index/h/honey-barbequed-salmon-with-mache-12667.json)
* [Honey Bourbon Toddy 368975](../../index/h/honey-bourbon-toddy-368975.json)
* [Honey Brandy Oranges 107171](../../index/h/honey-brandy-oranges-107171.json)
* [Honey Brined Turkey With Giblet Cream Gravy 102478](../../index/h/honey-brined-turkey-with-giblet-cream-gravy-102478.json)
* [Honey Butter French Bread 102470](../../index/h/honey-butter-french-bread-102470.json)
* [Honey Caramel Tart With Apricots And Almonds 237924](../../index/h/honey-caramel-tart-with-apricots-and-almonds-237924.json)
* [Honey Cardamom Custards With Strawberry Orange Compote 1580](../../index/h/honey-cardamom-custards-with-strawberry-orange-compote-1580.json)
* [Honey Cheese Pie With Strawberry Compote 1705](../../index/h/honey-cheese-pie-with-strawberry-compote-1705.json)
* [Honey Cinnamon Ice Cream 364329](../../index/h/honey-cinnamon-ice-cream-364329.json)
* [Honey Citrus Chicken 2280](../../index/h/honey-citrus-chicken-2280.json)
* [Honey Custard In Phyllo Pastry 1070](../../index/h/honey-custard-in-phyllo-pastry-1070.json)
* [Honey Date And Pecan Tart 241628](../../index/h/honey-date-and-pecan-tart-241628.json)
* [Honey Date Nut Breads 11080](../../index/h/honey-date-nut-breads-11080.json)
* [Honey Dew And Sake Granita 12376](../../index/h/honey-dew-and-sake-granita-12376.json)
* [Honey Eggnog 200915](../../index/h/honey-eggnog-200915.json)
* [Honey Ginger Barbecue Sauce 51108020](../../index/h/honey-ginger-barbecue-sauce-51108020.json)
* [Honey Ginger Pineapple Crepes 103208](../../index/h/honey-ginger-pineapple-crepes-103208.json)
* [Honey Ginger Salmon 5908](../../index/h/honey-ginger-salmon-5908.json)
* [Honey Gingered Pork Tenderloins 15277](../../index/h/honey-gingered-pork-tenderloins-15277.json)
* [Honey Glazed Baked Apples Filled With Dried Fruit 102520](../../index/h/honey-glazed-baked-apples-filled-with-dried-fruit-102520.json)
* [Honey Glazed Butternut Squash 4580](../../index/h/honey-glazed-butternut-squash-4580.json)
* [Honey Glazed Carrots With Carrot Top Gremolata](../../index/h/honey-glazed-carrots-with-carrot-top-gremolata.json)
* [Honey Glazed Doughnuts With Raisins And Pine Nuts 236796](../../index/h/honey-glazed-doughnuts-with-raisins-and-pine-nuts-236796.json)
* [Honey Glazed Duck With Fig And Pistachio Red Quinoa 51125200](../../index/h/honey-glazed-duck-with-fig-and-pistachio-red-quinoa-51125200.json)
* [Honey Glazed Grilled Pineapple 108252](../../index/h/honey-glazed-grilled-pineapple-108252.json)
* [Honey Glazed Lemon Chicken 109394](../../index/h/honey-glazed-lemon-chicken-109394.json)
* [Honey Glazed Oven Roasted Sweet Potato Wedges 236503](../../index/h/honey-glazed-oven-roasted-sweet-potato-wedges-236503.json)
* [Honey Glazed Peach Tart With Mascarpone Cream 108473](../../index/h/honey-glazed-peach-tart-with-mascarpone-cream-108473.json)
* [Honey Glazed Rack Of Lamb 4392](../../index/h/honey-glazed-rack-of-lamb-4392.json)
* [Honey Glazed Vegetables](../../index/h/honey-glazed-vegetables.json)
* [Honey Hazelnut Financiers 51234850](../../index/h/honey-hazelnut-financiers-51234850.json)
* [Honey Lemon Buttermilk Dressing 12083](../../index/h/honey-lemon-buttermilk-dressing-12083.json)
* [Honey Lemon Custard With Fruit 51104290](../../index/h/honey-lemon-custard-with-fruit-51104290.json)
* [Honey Mustard And Rosemary Pork Roast 2130](../../index/h/honey-mustard-and-rosemary-pork-roast-2130.json)
* [Honey Mustard Chicken Sausage Kebabs 353417](../../index/h/honey-mustard-chicken-sausage-kebabs-353417.json)
* [Honey Mustard Dressing 11578](../../index/h/honey-mustard-dressing-11578.json)
* [Honey Mustard Dressing And Mixed Greens 104167](../../index/h/honey-mustard-dressing-and-mixed-greens-104167.json)
* [Honey Mustard Glazed Shrimp And Scallops 238009](../../index/h/honey-mustard-glazed-shrimp-and-scallops-238009.json)
* [Honey Mustard Sauce 109036](../../index/h/honey-mustard-sauce-109036.json)
* [Honey Mustard Sauce 5442](../../index/h/honey-mustard-sauce-5442.json)
* [Honey Nut Squares 107829](../../index/h/honey-nut-squares-107829.json)
* [Honey Oat Bread 105775](../../index/h/honey-oat-bread-105775.json)
* [Honey Orange Butternut Squash 10650](../../index/h/honey-orange-butternut-squash-10650.json)
* [Honey Orange Madeleines 1807](../../index/h/honey-orange-madeleines-1807.json)
* [Honey Pan Dulce With Nuts And Dried Fruit 236470](../../index/h/honey-pan-dulce-with-nuts-and-dried-fruit-236470.json)
* [Honey Poached Pears With Mascarpone 100535](../../index/h/honey-poached-pears-with-mascarpone-100535.json)
* [Honey Roasted Carrots With Tahini Yogurt 51254840](../../index/h/honey-roasted-carrots-with-tahini-yogurt-51254840.json)
* [Honey Roasted Cherry Tomatoes 51174600](../../index/h/honey-roasted-cherry-tomatoes-51174600.json)
* [Honey Roasted Chicken With Lemon And Tarragon 108489](../../index/h/honey-roasted-chicken-with-lemon-and-tarragon-108489.json)
* [Honey Roasted Ham Or Turkey With Dried Cherry Relish 5774](../../index/h/honey-roasted-ham-or-turkey-with-dried-cherry-relish-5774.json)
* [Honey Roasted Lamb With Arugula And Pine Nut Salad 106475](../../index/h/honey-roasted-lamb-with-arugula-and-pine-nut-salad-106475.json)
* [Honey Roasted Onion Tart 363711](../../index/h/honey-roasted-onion-tart-363711.json)
* [Honey Roasted Pear Salad With Thyme Verjus Dressing 232870](../../index/h/honey-roasted-pear-salad-with-thyme-verjus-dressing-232870.json)
* [Honey Rosemary Ice Cream 363713](../../index/h/honey-rosemary-ice-cream-363713.json)
* [Honey Rye 367137](../../index/h/honey-rye-367137.json)
* [Honey Tree Bark Biscuits 243193](../../index/h/honey-tree-bark-biscuits-243193.json)
* [Honey Turmeric Pork With Beet And Carrot Salad 56389766](../../index/h/honey-turmeric-pork-with-beet-and-carrot-salad-56389766.json)
* [Honey Vanilla Poached Apricots 12387](../../index/h/honey-vanilla-poached-apricots-12387.json)
* [Honey Vinegar Leg Of Lamb With Fennel And Carrots 51193690](../../index/h/honey-vinegar-leg-of-lamb-with-fennel-and-carrots-51193690.json)
* [Honey Yogurt 354340](../../index/h/honey-yogurt-354340.json)
* [Jalapeno Honey Mustard Pork Kebabs 10142](../../index/j/jalapeno-honey-mustard-pork-kebabs-10142.json)
* [Just As Sweet As Honey Cake 369050](../../index/j/just-as-sweet-as-honey-cake-369050.json)
* [Lavender Honey And Yogurt Pie 238748](../../index/l/lavender-honey-and-yogurt-pie-238748.json)
* [Lavender Honey Tea Bread 233911](../../index/l/lavender-honey-tea-bread-233911.json)
* [Lemon Honey Almond Cake 101412](../../index/l/lemon-honey-almond-cake-101412.json)
* [Lemon Honey And Walnut Tart 1473](../../index/l/lemon-honey-and-walnut-tart-1473.json)
* [Lemon Honey Tart With Salted Shortbread Crust 51140700](../../index/l/lemon-honey-tart-with-salted-shortbread-crust-51140700.json)
* [Lime And Honey Glazed Salmon With Basmati And Broccolini 355189](../../index/l/lime-and-honey-glazed-salmon-with-basmati-and-broccolini-355189.json)
* [Low Fat Yogurt Honey Mint Dressing 12219](../../index/l/low-fat-yogurt-honey-mint-dressing-12219.json)
* [Madeleines With Lavender Honey 357229](../../index/m/madeleines-with-lavender-honey-357229.json)
* [Maltese Chicken With Red Wine And Honey Gravy 106485](../../index/m/maltese-chicken-with-red-wine-and-honey-gravy-106485.json)
* [Marjoram Honey Vinaigrette 240032](../../index/m/marjoram-honey-vinaigrette-240032.json)
* [Mascarpone Tart With Honey Oranges And Pistachios 237299](../../index/m/mascarpone-tart-with-honey-oranges-and-pistachios-237299.json)
* [Mexican Chopped Salad With Honey Lime Dressing 230154](../../index/m/mexican-chopped-salad-with-honey-lime-dressing-230154.json)
* [Mixed Greens With Honey Mustard Dressing 11398](../../index/m/mixed-greens-with-honey-mustard-dressing-11398.json)
* [Mixed Greens With Sheeps Milk Cheese And Honey Vinaigrette 236189](../../index/m/mixed-greens-with-sheeps-milk-cheese-and-honey-vinaigrette-236189.json)
* [Molasses Honey Ginger Cookies 106791](../../index/m/molasses-honey-ginger-cookies-106791.json)
* [Nectarine Tarts With Honey Ice Cream 2330](../../index/n/nectarine-tarts-with-honey-ice-cream-2330.json)
* [Nectarines Strawberries And Melon In Orange Honey Spiced Syrup 5411](../../index/n/nectarines-strawberries-and-melon-in-orange-honey-spiced-syrup-5411.json)
* [Oat And Dried Fruit Cereal With Honey And Yogurt 4477](../../index/o/oat-and-dried-fruit-cereal-with-honey-and-yogurt-4477.json)
* [Old Fashioned Oatmeal With Apples Raisins And Honey Toasted Walnuts 3116](../../index/o/old-fashioned-oatmeal-with-apples-raisins-and-honey-toasted-walnuts-3116.json)
* [Orange Blossom Honey Madeleines 367764](../../index/o/orange-blossom-honey-madeleines-367764.json)
* [Orange Honey Sauce 103922](../../index/o/orange-honey-sauce-103922.json)
* [Orange Honey Syrup 10414](../../index/o/orange-honey-syrup-10414.json)
* [Orange Roasted Baby Carrots With Honey 237912](../../index/o/orange-roasted-baby-carrots-with-honey-237912.json)
* [Orange Spiced Rye Honey Cake Gefilte Manifesto Recipe](../../index/o/orange-spiced-rye-honey-cake-gefilte-manifesto-recipe.json)
* [Oranges With Pomegranate Molasses And Honey 350797](../../index/o/oranges-with-pomegranate-molasses-and-honey-350797.json)
* [Oven Poached Pears With Lemon Honey And Nutmeg 5160](../../index/o/oven-poached-pears-with-lemon-honey-and-nutmeg-5160.json)
* [Passover Honey Nut Cake In Soaking Syrup 109151](../../index/p/passover-honey-nut-cake-in-soaking-syrup-109151.json)
* [Pastry Twists With Spiced Sugar Honey Glaze 358569](../../index/p/pastry-twists-with-spiced-sugar-honey-glaze-358569.json)
* [Pear Cake With Lemon Honey Cream Cheese Frosting 104048](../../index/p/pear-cake-with-lemon-honey-cream-cheese-frosting-104048.json)
* [Pear Crostata With Figs And Honey 361795](../../index/p/pear-crostata-with-figs-and-honey-361795.json)
* [Pears In Honey And Pine Nut Caramel With Artisanal Cheese 241350](../../index/p/pears-in-honey-and-pine-nut-caramel-with-artisanal-cheese-241350.json)
* [Pears With Herbed Ricotta And Honey 232639](../../index/p/pears-with-herbed-ricotta-and-honey-232639.json)
* [Pecan Biscuits With Orange Honey Glaze 5014](../../index/p/pecan-biscuits-with-orange-honey-glaze-5014.json)
* [Pecan Waffles With Sauteed Bananas And Cinnamon Honey 363321](../../index/p/pecan-waffles-with-sauteed-bananas-and-cinnamon-honey-363321.json)
* [Phyllo Nests With Strawberries And Honey 108131](../../index/p/phyllo-nests-with-strawberries-and-honey-108131.json)
* [Pine Nut Honey Ice Cream 105233](../../index/p/pine-nut-honey-ice-cream-105233.json)
* [Plantain Flatbread With Poached Egg And Honey Drizzle](../../index/p/plantain-flatbread-with-poached-egg-and-honey-drizzle.json)
* [Plum Fennel Salad With Honey Ginger Dressing 56389860](../../index/p/plum-fennel-salad-with-honey-ginger-dressing-56389860.json)
* [Plum Tarts With Honey And Black Pepper 51239610](../../index/p/plum-tarts-with-honey-and-black-pepper-51239610.json)
* [Pork Kebabs Marinated In Honey Rosemary And Orange 106479](../../index/p/pork-kebabs-marinated-in-honey-rosemary-and-orange-106479.json)
* [Port And Honey Poached Pears With Lemon Curd Mousse 12712](../../index/p/port-and-honey-poached-pears-with-lemon-curd-mousse-12712.json)
* [Profiteroles With Honey Lavender Ice Cream 103082](../../index/p/profiteroles-with-honey-lavender-ice-cream-103082.json)
* [Puff Pastry Apple Tarts Glazed With Honey 107174](../../index/p/puff-pastry-apple-tarts-glazed-with-honey-107174.json)
* [Puff Pastry Baskets With Honey Mousse And Assorted Fruit 1904](../../index/p/puff-pastry-baskets-with-honey-mousse-and-assorted-fruit-1904.json)
* [Red Chili And Honey Glazed Turkey With Ancho Pan Gravy 104309](../../index/r/red-chili-and-honey-glazed-turkey-with-ancho-pan-gravy-104309.json)
* [Rice Pudding With Fresh Pears And Honey 51246410](../../index/r/rice-pudding-with-fresh-pears-and-honey-51246410.json)
* [Roast Chicken With Rosemary Lemon And Honey 395509](../../index/r/roast-chicken-with-rosemary-lemon-and-honey-395509.json)
* [Roast Chicken With Saffron Hazelnuts And Honey 51184640](../../index/r/roast-chicken-with-saffron-hazelnuts-and-honey-51184640.json)
* [Roasted Apricots With Honey Vanilla Creme Fraiche 359249](../../index/r/roasted-apricots-with-honey-vanilla-creme-fraiche-359249.json)
* [Roasted Honey Glazed Spareribs 100570](../../index/r/roasted-honey-glazed-spareribs-100570.json)
* [Roasted Hot Honey Shrimp With Bok Choy And Kimchi Rice 51261050](../../index/r/roasted-hot-honey-shrimp-with-bok-choy-and-kimchi-rice-51261050.json)
* [Roasted Parsnips Turnips And Rutabagas With Ancho Spiced Honey Glaze 361291](../../index/r/roasted-parsnips-turnips-and-rutabagas-with-ancho-spiced-honey-glaze-361291.json)
* [Roasted Pineapple With Honey And Pistachios 394677](../../index/r/roasted-pineapple-with-honey-and-pistachios-394677.json)
* [Roasted Spiced Chicken With Cinnamon And Honey Glazed Sweet Potatoes 234808](../../index/r/roasted-spiced-chicken-with-cinnamon-and-honey-glazed-sweet-potatoes-234808.json)
* [Roasted Sweet Potatoes With Honey Glaze 104728](../../index/r/roasted-sweet-potatoes-with-honey-glaze-104728.json)
* [Rustic Apple Tart With Honey Dates And Nuts 108599](../../index/r/rustic-apple-tart-with-honey-dates-and-nuts-108599.json)
* [Sage And Honey Skillet Cornbread 240376](../../index/s/sage-and-honey-skillet-cornbread-240376.json)
* [Salmon Glazed With Honey And Mustard 102649](../../index/s/salmon-glazed-with-honey-and-mustard-102649.json)
* [Salmon Glazed With Honey And Mustard 350653](../../index/s/salmon-glazed-with-honey-and-mustard-350653.json)
* [Salmon With Soy Honey And Wasabi Sauces 105026](../../index/s/salmon-with-soy-honey-and-wasabi-sauces-105026.json)
* [Scottish Salmon With Shallot Truffle Honey Glaze Lump Crab And Green Apple Risotto And Quince Jam 363590](../../index/s/scottish-salmon-with-shallot-truffle-honey-glaze-lump-crab-and-green-apple-risotto-and-quince-jam-363590.json)
* [Seared Duck Breast With Chili Honey Ginger Glaze 102303](../../index/s/seared-duck-breast-with-chili-honey-ginger-glaze-102303.json)
* [Seared Scallops With Spicy Honey Citrus Glaze On Celery Salad 239848](../../index/s/seared-scallops-with-spicy-honey-citrus-glaze-on-celery-salad-239848.json)
* [Skewered Grilled Fruit With Minted Yogurt Honey Sauce 105250](../../index/s/skewered-grilled-fruit-with-minted-yogurt-honey-sauce-105250.json)
* [Skewered Honey Glazed Scallops With Peach Salsa 109692](../../index/s/skewered-honey-glazed-scallops-with-peach-salsa-109692.json)
* [Smoked Baby Back Ribs With Hoisin Honey Glaze 105305](../../index/s/smoked-baby-back-ribs-with-hoisin-honey-glaze-105305.json)
* [Spiced Fresh Orange And Honey Sorbet 241111](../../index/s/spiced-fresh-orange-and-honey-sorbet-241111.json)
* [Spiced Honey Cake With Cream Cheese Frosting 56389999](../../index/s/spiced-honey-cake-with-cream-cheese-frosting-56389999.json)
* [Spicy Honey Glazed Parsnips 51223670](../../index/s/spicy-honey-glazed-parsnips-51223670.json)
* [Spicy Honey Mustard Sauce 56389343](../../index/s/spicy-honey-mustard-sauce-56389343.json)
* [Strawberry Panna Cotta With Strawberries And Honey 103528](../../index/s/strawberry-panna-cotta-with-strawberries-and-honey-103528.json)
* [Sweet Honey And Saffron Challah 361129](../../index/s/sweet-honey-and-saffron-challah-361129.json)
* [Sweet Potato And Poblano Salad With Honey And Rosemary 353851](../../index/s/sweet-potato-and-poblano-salad-with-honey-and-rosemary-353851.json)
* [Sweet Potato Biscuits With Ham Mustard And Honey 356312](../../index/s/sweet-potato-biscuits-with-ham-mustard-and-honey-356312.json)
* [Sweet Potato Pancakes With Honey Cinnamon Butter 239216](../../index/s/sweet-potato-pancakes-with-honey-cinnamon-butter-239216.json)
* [Sweet Potatoes And Greens With Honey Mustard Glaze 2605](../../index/s/sweet-potatoes-and-greens-with-honey-mustard-glaze-2605.json)
* [Swiss Honey Walnut Tart 4663](../../index/s/swiss-honey-walnut-tart-4663.json)
* [Tamarind Honey Lamb Kebabs On Mashed Yams 231084](../../index/t/tamarind-honey-lamb-kebabs-on-mashed-yams-231084.json)
* [Tangerine Honey Flan With Grapefruit Segments 234825](../../index/t/tangerine-honey-flan-with-grapefruit-segments-234825.json)
* [Tea And Honey Crisps 236667](../../index/t/tea-and-honey-crisps-236667.json)
* [Tricolor Salad With Honey Cumin Dressing 1738](../../index/t/tricolor-salad-with-honey-cumin-dressing-1738.json)
* [Turkey Cheddar Sandwiches With Honey Mustard 101442](../../index/t/turkey-cheddar-sandwiches-with-honey-mustard-101442.json)
* [Twice Roasted Sweet Potatoes With Hot Honey](../../index/t/twice-roasted-sweet-potatoes-with-hot-honey.json)
* [Upside Down Honey Cheesecakes 237917](../../index/u/upside-down-honey-cheesecakes-237917.json)
* [Vanilla Bean Roasted Figs With Wildflower Honey Vanilla Ice Cream 105862](../../index/v/vanilla-bean-roasted-figs-with-wildflower-honey-vanilla-ice-cream-105862.json)
* [Vidalia Onion Tart With Bacon Local Honey And Fresh Thyme 354950](../../index/v/vidalia-onion-tart-with-bacon-local-honey-and-fresh-thyme-354950.json)
* [Walnut Apricot Cake In Honey Syrup 106498](../../index/w/walnut-apricot-cake-in-honey-syrup-106498.json)
* [Walnut Thyme Honey 368990](../../index/w/walnut-thyme-honey-368990.json)
* [Warm Goat Cheese And Carrot Salad With Marjoram Honey Vinaigrette 240031](../../index/w/warm-goat-cheese-and-carrot-salad-with-marjoram-honey-vinaigrette-240031.json)
* [Warm Goat Cheese Toasts With Rosemary Walnuts And Honey 108799](../../index/w/warm-goat-cheese-toasts-with-rosemary-walnuts-and-honey-108799.json)
* [Whipped Ricotta With Honey And Mixed Berries 105374](../../index/w/whipped-ricotta-with-honey-and-mixed-berries-105374.json)
* [White Sweet Potatoes With Mirin And Honey 388730](../../index/w/white-sweet-potatoes-with-mirin-and-honey-388730.json)
* [Wildflower Honey Semifreddo With Honey Sesame Wafers 232643](../../index/w/wildflower-honey-semifreddo-with-honey-sesame-wafers-232643.json)
* [Wildflower Honey Semifreddo With Honey Sesame Wafers 232643](../../index/w/wildflower-honey-semifreddo-with-honey-sesame-wafers-232643.json)
* [Wilted Autumn Greens With Honey Mustard Vinaigrette 51211620](../../index/w/wilted-autumn-greens-with-honey-mustard-vinaigrette-51211620.json)
* [Yogurt With Fresh Figs Honey And Pine Nuts 51239840](../../index/y/yogurt-with-fresh-figs-honey-and-pine-nuts-51239840.json)
|
2024-11-18T18:08:23.225345+00:00 | 2023-08-31T02:42:05 | 771069790c39bc89d870d67c3a484207eb566c2d | 3 | {
"blob_id": "771069790c39bc89d870d67c3a484207eb566c2d",
"branch_name": "refs/heads/main",
"committer_date": "2023-08-31T02:42:05",
"content_id": "2ddd323ad7c5bfe50cfb38d5f2e038a1780134e3",
"detected_licenses": [
"BSD-3-Clause",
"MIT"
],
"directory_id": "ed24e9cd5af2c77f6c59d7b31c9478a1d9d2fbf2",
"extension": "md",
"filename": "CONTRIBUTING.md",
"fork_events_count": 1,
"gha_created_at": "2021-08-03T19:42:38",
"gha_event_created_at": "2023-08-31T02:42:06",
"gha_language": "Python",
"gha_license_id": "MIT",
"github_id": 392435742,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 824,
"license": "BSD-3-Clause,MIT",
"license_type": "permissive",
"path": "/CONTRIBUTING.md",
"provenance": "stack-edu-markdown-0000.json.gz:8",
"repo_name": "decargroup/pykoop",
"revision_date": "2023-08-31T02:42:05",
"revision_id": "77a7a0d8fe63f286cf2d587af41278658d215552",
"snapshot_id": "e47a97e224b6e2ce9b2c14136f94f01a69cde3e8",
"src_encoding": "UTF-8",
"star_events_count": 20,
"url": "https://raw.githubusercontent.com/decargroup/pykoop/77a7a0d8fe63f286cf2d587af41278658d215552/CONTRIBUTING.md",
"visit_date": "2023-09-01T20:11:08.668788"
} | 3.296875 | stackv2 | # Contribution Guidelines
Everyone is welcome to contribute to `pykoop`. To do so, please fork the
repository and create a pull request when you're ready.
Contributed code must
1. be documented using
[NumPy style docstrings](https://sphinxcontrib-napoleon.readthedocs.io/en/latest/example_numpy.html),
2. be formatted using using
[YAPF](https://github.com/google/yapf) (see `./.style.yapf`),
3. use [type annotations](https://docs.python.org/3/library/typing.html)
consistently,
4. include relevant unit tests, and
5. pass existing unit tests and checks.
If you are fixing a bug, please include a set of unit tests in your pull
request that would fail without your changes.
If you notice a problem or would like to suggest an enhancement, please create
an [issue](https://github.com/decargroup/pykoop/issues).
|
2024-11-18T18:08:23.295125+00:00 | 2021-09-24T14:45:32 | 1bb913214dbf3debdca15253188a418a2b70d57f | 3 | {
"blob_id": "1bb913214dbf3debdca15253188a418a2b70d57f",
"branch_name": "refs/heads/master",
"committer_date": "2021-09-24T14:45:32",
"content_id": "0db6af5e56aa45691b0ce6c4fe599d7085bc10c8",
"detected_licenses": [
"MIT"
],
"directory_id": "9c7b9ba5ad36055b144f9d215498f1d596f49365",
"extension": "md",
"filename": "2021-06-28-Post00035.md",
"fork_events_count": 0,
"gha_created_at": "2021-03-27T19:32:03",
"gha_event_created_at": "2021-03-27T20:30:41",
"gha_language": "SCSS",
"gha_license_id": "MIT",
"github_id": 352159974,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 5357,
"license": "MIT",
"license_type": "permissive",
"path": "/_posts/2021-06-28-Post00035.md",
"provenance": "stack-edu-markdown-0000.json.gz:9",
"repo_name": "laifukun/laifukun.github.io",
"revision_date": "2021-09-24T14:45:32",
"revision_id": "e3beacdf34d61f58d3a8fb70c48110a18c91976b",
"snapshot_id": "a2bfb7858f8b524f142a49ef845cc65225668362",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/laifukun/laifukun.github.io/e3beacdf34d61f58d3a8fb70c48110a18c91976b/_posts/2021-06-28-Post00035.md",
"visit_date": "2023-08-05T08:18:45.444795"
} | 3.234375 | stackv2 | ---
layout: post
title: "Algorithm & Data Structure - Substring with Bit Mask"
---
A number of substring problems could be solved by bit mask with conversion to similar as subarray sum problems. Basic subarray sum equals K could be soved by hashmap in O(n):
1. Store number of prefix sum in hashmap.
2. Get the number of current running sum - K.
* LeetCode - 560
```java
public int subarraySum(int[] nums, int k) {
int count = 0, sum = 0;
HashMap<Integer, Integer> sumMap = new HashMap<Integer, Integer>();
sumMap.put(sum,1);
for (int i = 0; i < nums.length; i++) {
sum += nums[i];
if (sumMap.containsKey(sum - k))
count += sumMap.get(sum-k);
sumMap.put(sum, sumMap.getOrDefault(sum, 0)+1);
}
return count;
}
```
Similary approach could be applied to a few substring problem with the help of bitmask. Bitmask is able to convert the substring to a number, and store it in the hashmap, especially if the problem requires to find even or odd number of characters.
* LeetCode - 1371
* LeetCode - 1542
* LeetCode - 1915

Because we only care even or odd number of characters, we could use bitmask to reverse the bit with 0 or 1, i.e., if the number of character is even, set it to 0, else set it to 1. A bit operation XOR could achieve it easily. We only need to XOR the bit if we see the corresponding character. Because the string only include lowercase letter, we could fit all the letters to a 32 bit integer.
We reverse its bit whenever we see a vowel. If two positions $$i$$ and $$j$$ have the see bit pattern (mask), that means, substring [i+1....j] has even number of vowels. Why? because if a vowel's bit is 0 at both positions, this means from [0...i] and [0...j], the letter appears even times, so it shoud also be even times from [i+1...j], (even number - even number = even number). If a vowwel's bit is 1 at both positions, this means from [0...i] and [0...j], the letter appears odd times, so it shoud also be odd times from [i+1...j], (odd number - odd number = even number).
Since we are looking for the max length, we only need to store the first time a mask appears, and for a mask of 0, it means it is position is -1.
```java
public int findTheLongestSubstring(String s) {
int mask = 0;
Map<Integer, Integer> vowelCount = new HashMap<>();
vowelCount.put(0,-1);
int length = 0;
for (int i = 0; i < s.length(); i++) {
char c = s.charAt(i);
if (c == 'a' || c=='e' || c=='i' || c=='o' || c=='u'){
mask ^= (1<<(c-'a'));
}
if (!vowelCount.containsKey(mask)) {
vowelCount.put(mask, i);
} else {
length = Math.max(length, i - vowelCount.get(mask));
}
}
return length;
}
```

We use the similar bit mask approach to represent odd or even numbers of each character. We have only 10 digits, so it could fit in a 32 bit integer. We use the same to revers a bit and 0 represents even time, 1 represents odd times so far. Also we only need to record the first time position of a mask.
The biggest difference here is we are not only looking for the position of the same mask, but also the position of a mask with one letter difference. Because, palindrome string could also be formed with an addition letter. How do we achieve this, we could find the mask with one bit different from the map.
```java
public int longestAwesome(String s) {
int mask = 0;
Map<Integer, Integer> digitCount = new HashMap<>();
digitCount.put(0, -1);
int length = 0;
for (int i = 0; i < s.length(); i++) {
mask ^= (1<<(s.charAt(i)-'0'));
//find the first position with the same mask
length = Math.max(length, i - digitCount.getOrDefault(mask, i);
// find the first position with one bit difference
for (int k = 0; k < 10; k++) {
length = Math.max(length, i - digitCount.getOrDefault(mask^(1<<k), i));
}
if (!digitCount.containsKey(mask))
digitCount.put(mask, i);
}
return length;
}
```

This is more close to Leetcode-1542 because the similar 10 letters and allow one odd difference. But this problem requires the record the number of time a mask appeared. So it is similar to subarray sum equals K problem above.
```java
public long wonderfulSubstrings(String word) {
Map<Integer, Integer> maskCount = new HashMap<>();
maskCount.put(0,1);
int mask = 0;
long count = 0;
for (int i = 0; i < word.length(); i++) {
char c = word.charAt(i);
mask ^= 1<<(c-'a');
count += maskCount.getOrDefault(mask,0);
for (int j = 0; j < 10; j++) {
count += maskCount.getOrDefault(mask^(1<<j),0);
}
maskCount.put(mask, maskCount.getOrDefault(mask,0)+1);
}
return count;
}
```
|
2024-11-18T18:08:23.475761+00:00 | 2020-11-06T10:29:20 | f32ceae0ea0459dd07a63935aa9ab1493d289dee | 3 | {
"blob_id": "f32ceae0ea0459dd07a63935aa9ab1493d289dee",
"branch_name": "refs/heads/develop",
"committer_date": "2020-11-06T10:29:20",
"content_id": "4fb5b8c99410b5409864ad27bb9a56a40eb5ae03",
"detected_licenses": [
"Apache-2.0"
],
"directory_id": "e24346204dd7c88bd143e9604e5af7ff40fb5c7a",
"extension": "md",
"filename": "AUTHORS.md",
"fork_events_count": 0,
"gha_created_at": "2020-11-07T04:51:57",
"gha_event_created_at": "2020-12-04T15:58:44",
"gha_language": null,
"gha_license_id": "Apache-2.0",
"github_id": 310769850,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 1692,
"license": "Apache-2.0",
"license_type": "permissive",
"path": "/AUTHORS.md",
"provenance": "stack-edu-markdown-0000.json.gz:10",
"repo_name": "crtn32002/element-android",
"revision_date": "2020-11-06T10:29:20",
"revision_id": "0753ba3495e8120c2667b7bcc97904c23cd8819f",
"snapshot_id": "f294e81c206ec354998d47bcda8ea42ba5d52076",
"src_encoding": "UTF-8",
"star_events_count": 1,
"url": "https://raw.githubusercontent.com/crtn32002/element-android/0753ba3495e8120c2667b7bcc97904c23cd8819f/AUTHORS.md",
"visit_date": "2023-01-19T09:11:54.611276"
} | 2.8125 | stackv2 | A full developer contributors list can be found [here](https://github.com/vector-im/element-android/graphs/contributors).
# Core team:
Even if we try to be able to work on all the functionalities, we have more knowledge about what we have developed ourselves.
## Benoit: Android team leader
[@benoit.marty:matrix.org](https://matrix.to/#/@benoit.marty:matrix.org)
- Android team leader and project leader, Android developer, GitHub community manager.
- Specialist of the account creation, and many other fun features.
- Reviewing and polishing developed features, code quality manager, PRs reviewer, GitHub community manager.
- Release manager on the Play Store
## François: Software architect
[@ganfra:matrix.org](https://matrix.to/#/@ganfra:matrix.org)
- Software architect, Android developer
- First developer on the project.
- Work mainly on the global architecture of the project.
- Specialist of the timeline, and lots of other features.
## Valere: Product manager, Android developer
[@valere35:matrix.org](https://matrix.to/#/@valere35:matrix.org)
- Product manager, Android developer
- Specialist on the crypto implementation.
# Other contributors
First of all, we thank all contributors who use Element and report problems on this GitHub project or via the integrated rageshake function.
We do not forget all translators, for their work of translating Element into many languages. They are also the authors of Element.
Feel free to add your name below, when you contribute to the project!
Name | Matrix ID | GitHub
--------|---------------------|--------------------------------------
gjpower | @gjpower:matrix.org | [gjpower](https://github.com/gjpower)
|
2024-11-18T18:08:23.629672+00:00 | 2021-10-04T03:48:56 | 419878f65064585ce4bc38482f1d46182694f25d | 4 | {
"blob_id": "419878f65064585ce4bc38482f1d46182694f25d",
"branch_name": "refs/heads/master",
"committer_date": "2021-10-04T03:48:56",
"content_id": "a2dc65dd4de09662f9b66990969cff1c3fe6cd91",
"detected_licenses": [
"MIT"
],
"directory_id": "4b7b8656598e32182d80222d428e0118049067e1",
"extension": "md",
"filename": "index.md",
"fork_events_count": 0,
"gha_created_at": "2020-09-30T19:17:18",
"gha_event_created_at": "2020-09-30T19:17:19",
"gha_language": null,
"gha_license_id": "MIT",
"github_id": 300033772,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 3979,
"license": "MIT",
"license_type": "permissive",
"path": "/content/react-constructor-get-initial-state/index.md",
"provenance": "stack-edu-markdown-0000.json.gz:11",
"repo_name": "iamdarshshah/engineering-portal",
"revision_date": "2021-10-04T03:48:56",
"revision_id": "1e39af0cbc6900688fdb8ee6c5cb4fcf45c13934",
"snapshot_id": "a969b516c403d597d7e16331066a86e9335d6e6a",
"src_encoding": "UTF-8",
"star_events_count": 1,
"url": "https://raw.githubusercontent.com/iamdarshshah/engineering-portal/1e39af0cbc6900688fdb8ee6c5cb4fcf45c13934/content/react-constructor-get-initial-state/index.md",
"visit_date": "2023-08-01T16:56:10.723107"
} | 3.84375 | stackv2 | ---
title: "Constructor vs getInitialState in React"
date: "2021-01-05"
coverImage: "index.png"
author: Nathan Nguyen
tags: ["JavaScript", "React"]
description: "No ES6? No problem. getInitialState is the ES5 friendly method to define the initial state of a React component."
---
One fairly popular question that got asked on programming bulletin boards has to do with the similarities and differences between React’s <code>constructor</code> and the built in method <code>getInitialState</code>. While the simple answer to this question is indeed simple: “getInitialState is the ES5 friendly method to define the initial state of a React component,” there are a couple more details around <code>getInitialState</code> as well as React’s ES5 support that are interesting and useful to highlight.
## Constructor vs getInitialState: With or without Classes:
One of the fundamental differences between ES5 and ES6 in regards to React implementation is the new <code>class</code> keyword. It allows definition of React components as classes, which is a familiar data structure for anyone who has had experience with more traditional object-oriented languages such as Java or C++. The class structure also allows for natural organization of the component’s elements like state, props, lifecycle methods and member functions. However, ES5 did not provide the same convenience. So instead of defining a React component as a class:
```JavaScript
class HelloWorld extends React.Component {
render() {
return <span>Hello World</span>;
}
}
```
You would rely on a helper module called `create-react-class`:
```JavaScript
var createReactClass = require('create-react-class');
var HelloWorld = createReactClass({
render: function() {
return <span>Hello World</span>;
}
});
```
And it is within the object passed into `create-react-class` that you could define an initial state by populating the `getInitialState` attribute:
```JavaScript
var HelloWorld = createReactClass({
getInitialState: function() {
return {name: this.props.name};
},
render: function() {
return <span>Hello {this.state.name}</span>;
}
});
```
Which, in ES6 implementation would be the equivalent of:
```JavaScript
class HelloWorld extends React.Component {
constructor(props) {
super(props);
this.state = {
name: props.name
}
}
render() {
return <span>Hello {this.state.name}</span>;
}
}
```
## Autobinding
One difference worth noting is that the `create-react-class` method automatically binds `this` to every attribute method. This no longer holds true if you define React components using the common ES6 class syntax, making it so that you have to manually bind `this` to internal methods:
```JavaScript
class HelloWorld extends React.Component {
constructor(props) {
super(props);
this.state = {
name: props.name
}
this.changeName = this.changeName.bind(this);
}
changeName() {
this.setState({ name: "World" });
}
...
}
```
Or otherwise use the “arrow function” shorthand which takes care of binding:
```JavaScript
class HelloWorld extends React.Component {
constructor(props) {
super(props);
this.state = {
name: props.name
}
}
changeName = () => {
this.setState({ name: "World" });
}
...
}
```
## Parting Words
Since the update to ES6, there have been multiple new React iterations. You could now forgo the `constructor` declaration altogether and just declare `state` inline as a class member, or utilize React Hooks as a new way to initialize states. However, the ES5 support remains useful for legacy systems and adds to the overall flexibility of React as a toolset.
You can read more about React's ES5 support [in the official doc entry here](https://reactjs.org/docs/react-without-es6.html), and [the v0.13.0 beta release blog entry here](https://reactjs.org/blog/2015/01/27/react-v0.13.0-beta-1.html), for the respective ES6 changes.
|
2024-11-18T18:08:23.723873+00:00 | 2023-01-10T15:09:28 | 7e17d3702cd5fdec7f54d272ed4104eaea060e5a | 2 | {
"blob_id": "7e17d3702cd5fdec7f54d272ed4104eaea060e5a",
"branch_name": "refs/heads/master",
"committer_date": "2023-01-10T15:09:28",
"content_id": "78ce7576081b19b09b8661735a68f6a8b0e62234",
"detected_licenses": [
"MIT"
],
"directory_id": "53e3426fbdc30ff470d27f4f78462e7e3dad5eb9",
"extension": "md",
"filename": "what-every-programmer-should-know-about-SEO.md",
"fork_events_count": 0,
"gha_created_at": "2018-09-24T15:22:17",
"gha_event_created_at": "2019-07-20T21:11:01",
"gha_language": "HTML",
"gha_license_id": null,
"github_id": 150124062,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 1059,
"license": "MIT",
"license_type": "permissive",
"path": "/general-programming/seo/what-every-programmer-should-know-about-SEO.md",
"provenance": "stack-edu-markdown-0000.json.gz:12",
"repo_name": "eginwong/course-notes",
"revision_date": "2023-01-10T15:09:28",
"revision_id": "5a8a62d7f69d2e164ee80b11cbf18eaecb5130e1",
"snapshot_id": "722221a4318b929b79f0d35c078669b510354921",
"src_encoding": "UTF-8",
"star_events_count": 1,
"url": "https://raw.githubusercontent.com/eginwong/course-notes/5a8a62d7f69d2e164ee80b11cbf18eaecb5130e1/general-programming/seo/what-every-programmer-should-know-about-SEO.md",
"visit_date": "2023-02-01T02:24:22.527048"
} | 2.5 | stackv2 | # What Every Programmer Should Know About SEO
[ref](https://katemats.com/what-every-programmer-should-know-about-seo/)
- how do search engines work?
- user inputs query > query is categorized > scans collection of documents organized in inverted index > results are returned to user based on relevancy
- web crawler scrapes all web sites for metadata and contents and indexes collection
- relevancy = # query keywords * authority (pageRank, popularity)
- make sure your site is crawlable
- ensure pages render without JS enabled
- use descriptive anchor text to pages
- limit the # of links on the page
- follow URL best practices
- use the right keywords in all the right places
- put them in the URL
- title of page
- h1 tag
- alt text on images and descriptive image file names
- have contexually relevant text content on the page
- order of words matters
- don't spam
- avoid duplicate content
- use smart meta descriptions
- freshness is good
- fast site speed
- picking keywords
- building links
- references, social media |
2024-11-18T18:08:24.367050+00:00 | 2020-02-28T02:14:46 | b09f4d55d8aed2b847488530d0ce4445b69f8758 | 4 | {
"blob_id": "b09f4d55d8aed2b847488530d0ce4445b69f8758",
"branch_name": "refs/heads/master",
"committer_date": "2020-02-28T02:14:46",
"content_id": "7dcf0fe884c3f86e4759d03ad840c7d00b9cad55",
"detected_licenses": [
"Apache-2.0"
],
"directory_id": "b1eaf16db41e543964619d9cfec7a12055d290ab",
"extension": "md",
"filename": "README.md",
"fork_events_count": 1,
"gha_created_at": "2020-02-21T21:57:25",
"gha_event_created_at": "2020-04-02T02:12:27",
"gha_language": "Java",
"gha_license_id": "NOASSERTION",
"github_id": 242237783,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 2701,
"license": "Apache-2.0",
"license_type": "permissive",
"path": "/ion-kotlin/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:13",
"repo_name": "O330oei/BBOT",
"revision_date": "2020-02-28T02:14:46",
"revision_id": "c4a75c44f6e93d09ec44750b70a1af8549b5a2d0",
"snapshot_id": "ad9d6a8c42a7e7e8d82f9c82f1a6b8c6ee16c599",
"src_encoding": "UTF-8",
"star_events_count": 1,
"url": "https://raw.githubusercontent.com/O330oei/BBOT/c4a75c44f6e93d09ec44750b70a1af8549b5a2d0/ion-kotlin/README.md",
"visit_date": "2021-01-09T03:56:32.773381"
} | 3.765625 | stackv2 | # Ion Kotlin Extensions - async/await
async/await allows you to write code that looks like synchronous code, but is actually run asynchronously via coroutines.
For example, if you wanted to download a list of files with async/await in Ion:
```kotlin
fun getFiles(files: Array<String>) = async {
for (file in files) {
Ion.with(context)
.load(file)
.asString()
.await()
}
}
```
This may look like synchronous code, but it is not. The return type of getFiles is a actually [Future](https://github.com/koush/ion#futures). The operation happens asynchronously, and only when all the files are finished downloading, will the Future's callback be called.
The code in an async block is a [suspend fun, aka a coroutine](https://kotlinlang.org/docs/reference/coroutines.html).
```kotlin
async {
// the code in here is a suspend fun, a coroutine.
// execution can be suspended and resumed.
}
```
Inside an async block, you can use await on Futures to wait for their results.
```kotlin
fun getFiles(files: File) = async {
System.out.println("Bob")
val string = Ion.with(context)
.load(file)
.asString()
.await()
System.out.println("Chuck")
}
System.out.println("Alice")
getFiles(file)
System.out.println("David")
```
async blocks are run asynchronously, but look synchronous. The output from this code would be:
```
Alice
Bob
David
Chuck
```
Execution was paused at the await() call, and resumed after the file was downloaded.
Await can also be used to switch thread affinities:
```kotlin
async {
looper.await()
System.out.println("I'm running on this Looper thread.")
handler.await()
System.out.println("I'm running on this handler's Looper thread.")
executor.await()
System.out.println("I'm running on this Executor's thread pool.")
asyncServer.await()
System.out.println("I'm running on this AsyncServer's reactor thread.")
}
```
### Getting results from async functions
async lets you easily write functions that return Futures.
```kotlin
fun myStringFunction(url: String) = async {
try {
return@async Ion.with(context)
.load(url)
.asString()
.await()
}
catch (e: Exception) {
return@async "Failed to load"
}
}
```
myStringFunction returns a Future<String>. Can attach setCallback to get the result, or use it in other async code blocks
Or for brevity:
```kotlin
fun myStringFunction(url: String) = async {
try {
Ion.with(context)
.load(url)
.asString()
.await()
}
catch (e: Exception) {
"Failed to load"
}
}
```
|
2024-11-18T18:20:10.401501+00:00 | 2019-04-12T11:32:05 | 8dd402db35edc0fd81b65ace6074cd847c6c73db | 3 | {
"blob_id": "8dd402db35edc0fd81b65ace6074cd847c6c73db",
"branch_name": "refs/heads/master",
"committer_date": "2019-04-12T11:32:05",
"content_id": "95d004297897682c1c476f132ee6f66f88b7bc8e",
"detected_licenses": [
"MIT"
],
"directory_id": "27a73a25b153c826cb1378b05b29cdf44299ec02",
"extension": "md",
"filename": "readme.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": null,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 3705,
"license": "MIT",
"license_type": "permissive",
"path": "/readme.md",
"provenance": "stack-edu-markdown-0000.json.gz:14",
"repo_name": "ionic4-toolbox/App-MealDateFrontEnd-20190412",
"revision_date": "2019-04-12T11:32:05",
"revision_id": "486f1e8f92bdd5da89f547fb849901b416932b6d",
"snapshot_id": "6e1e0a8ec0a7c64b9213288e579680f0c06bc7c3",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/ionic4-toolbox/App-MealDateFrontEnd-20190412/486f1e8f92bdd5da89f547fb849901b416932b6d/readme.md",
"visit_date": "2020-05-09T11:04:38.727494"
} | 3.09375 | stackv2 | # MeelDate
> MeelDate is a social app that helps students around campus(or potentially everybody) find a partner to have a meal together and expand their social network.
This repo is still under development… this frontend project alone with our [**Backend Service**](https://github.com/xiaoxiao33/recommendation) are expected to be delivered to *Apple App Store && Google Play* on **April 27, 2019**.
## Demo(test-build-04090335)

To invite somebody out, just select the time and location, also note if you select "eat now", we'll emphasize the distance for recommendation, if you select "eat later", we will use a plethora of other informations including your gender, college.. and some behavior records like the people you like to provider a better and more personlized recommendation.
You can also check out our [**Video Demo**](https://youtu.be/mXjbKRyk4Zs), for Chinese visitors please go [**Here**](https://showmore.com/zh/u/mwk2n6s) for video. There're lots more to be discovered! Join our internal test group by just [sending us a<EMAIL_ADDRESS>if you are interested!
## Our App, You Decide
We'll be really appreciated if you can help us improve this app by [**leaving us your valuable feedback**](https://docs.google.com/forms/d/e/1FAIpQLSewlLvO2ZD1CiCeyDEw006TvVNfwBP0HOwXAutI-XjXQj-BkQ/viewform?vc=0&c=0&w=1), (only have 10 short choice questions)😄☺️😊🙏🤣😙😚🤑
## Quick Install
**ONLY for Registered Developers/Testers**

## Update
```
2019-04-12 07:31:00: add Yelp API
2019-04-09 01:26:00: add OTA distribution
2019-04-08 03:55:00: add animations
2019-04-05 20:25:00: UI improvements
2019-03-24 20:08:00: add GoogleMaps
2019-03-23 23:15:00: add invitation list page
2019-03-21 20:20:00: add recommendation page
2019-03-10 21:56:00: finish settings page
2019-03-09 23:48:00: FB login temporarily disabled
2019-03-09 22:53:00: allow facebook login
2019-03-09 17:28:00: pure bugs fix..
2019-02-25 13:54:00: add login authentication
2019-02-24 00:32:00: add static login page
2019-02-23 01:00:00: app created!
```
## Setup
#### 1, Node.js
```bash
# Check if you have `node.js`
# ([download_link](https://nodejs.org/en/download/))
node --version
```
#### 2, Ionic & Angular
```bash
# This comes with all of the Ionic components
# and Angular specific services and features.
npm install -g ionic
npm install -g @angular/cli
npm install @ionic/angular@latest --save
```
#### 3, Cordova(iOS Native Support)
```bash
# First, download and `cd` to this project
# then run the following command
npm i -g cordova
npm install -g ios-sim
npm install -g ios-deploy
ionic cordova prepare ios
```
## Build & Run
#### 1, iOS Emulation
```bash
# mac only, xcode required
# no need to add build&test params, I already include these in
# local script 'package.json.scripts'
npm run emulate-live
```
#### 2, Web Emulation
**WARNING:** this will not be supported in future due to the presence of native calls.
```bash
# -l stands for live reload
# -w=c means run with chrome
ionic serve -l -w=c
# open original
http://localhost:8100
```
#### 3, iOS Device
```bash
# Firt Run this to compile
ionic cordova prepare ios
# 1, Open Xcode. Use File » Open and locate the app. Open the app platforms/ios directory
# 2, Plug in your iphone into your PC
# 3, Select Build » and select your device to run the app!
```
#### 4, Android Device
Currently not supported(because none of our developers uses an android…) |
2024-11-18T18:20:10.569973+00:00 | 2022-01-13T14:30:48 | 538910c978aa9c6ef143e148f6d6141224b3d3e3 | 4 | {
"blob_id": "538910c978aa9c6ef143e148f6d6141224b3d3e3",
"branch_name": "refs/heads/master",
"committer_date": "2022-01-13T14:30:48",
"content_id": "6933442d5835425b8b2e3464ad020b9607c047d0",
"detected_licenses": [
"MIT"
],
"directory_id": "8da40867dc99de22ec58de1ee3b3e491cede2533",
"extension": "md",
"filename": "README.md",
"fork_events_count": 1,
"gha_created_at": "2019-06-27T13:13:28",
"gha_event_created_at": "2022-01-13T14:30:50",
"gha_language": "Java",
"gha_license_id": "MIT",
"github_id": 194098251,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 14338,
"license": "MIT",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:15",
"repo_name": "joeclark-phd/random-text-generators",
"revision_date": "2022-01-13T14:30:48",
"revision_id": "1d82a5675588bd67b1d9a4663de0076796c98d97",
"snapshot_id": "5f75bb7e4ee7de10fa13a2ed30e8bb6163c0a451",
"src_encoding": "UTF-8",
"star_events_count": 1,
"url": "https://raw.githubusercontent.com/joeclark-phd/random-text-generators/1d82a5675588bd67b1d9a4663de0076796c98d97/README.md",
"visit_date": "2022-01-15T01:45:45.013772"
} | 3.5 | stackv2 | # Random Text Generators by joeclark.net
[](https://search.maven.org/search?q=g:%22net.joeclark.proceduralgeneration%22%20AND%20a:%22randomtextgenerators%22)
[](https://github.com/joeclark-phd/random-text-generators/blob/master/LICENSE.md)
[](https://www.codacy.com/app/joeclark-phd/random-text-generators?utm_source=github.com&utm_medium=referral&utm_content=joeclark-phd/random-text-generators&utm_campaign=Badge_Grade)
[](https://lgtm.com/projects/g/joeclark-phd/random-text-generators/context:java)
API reference: [https://joeclark-phd.github.io/random-text-generators](https://joeclark-phd.github.io/random-text-generators/)
Example output: [https://github.com/joeclark-net/procedural-generation-examples](https://github.com/joeclark-phd/procedural-generation-examples)
This package defines an interface and implementation of a procedural random text generator that can be used, for example, to generate character or place names for an adventure game.
## Usage
Maven users, add this dependency to your POM:
<dependency>
<groupId>net.joeclark.proceduralgeneration</groupId>
<artifactId>randomtextgenerators</artifactId>
<version>1.0.1</version>
</dependency>
### RandomTextGenerator
The package offers an interface, **RandomTextGenerator**, with a single method:
- `String generateOne()` yields a new, procedurally-generated text string.
Currently there are five implementations of the interface:
- MarkovTextGenerator
- MarkovTextCasePreservingGenerator
- ClusterChainGenerator
- RandomDrawGenerator
- DoubleTextGenerator
### MarkovTextGenerator
Quick start:
RandomTextGenerator markov = new MarkovTextGenerator().train(myTextStream);
System.out.println(markov.generateOne());
(or with all the optional configuration...)
RandomTextGenerator markov = new MarkovTextGenerator().withOrder(2).withPrior(0.01).withStartFilter("J").withEndFilter("ia").withMinLength(3).withMaxLength(15).withRandom(myRandom).train(myTextStream);
System.out.println(markov.generateOne());
The big idea of Markov-chain random text generation is that you collect statistics on which characters follow other characters. So if a particular language uses "th" a lot, "t" should often be followed by "h" in the randomly-generated text. This class ingests a `Stream<String>` of training data to build up a Markov model, and uses it to generate new strings. However, the Markov-chain approach has a number of caveats:
First, looking only at two-character sequences isn't very sophisticated. The model would be smarter if you looked back more than one letter. For example, your model could know that "ot" and "nt" are often followed by "h" but "st" is not. The problem with that is that you will have far fewer examples of every 3-character, 4-character, or n-character sequences in your training data than you will have of 2-character sequences. If a sequence never occurs in your training data, it can never occur in your output. Because there are fewer examples, your output will be less random.
Based on an algorithm [described by JLund3 at RogueBasin](http://roguebasin.roguelikedevelopment.org/index.php?title=Names_from_a_high_order_Markov_Process_and_a_simplified_Katz_back-off_scheme), which I have also [implemented in Python](https://github.com/joeclark-phd/roguestate/blob/master/program/namegen.py), MarkovTextGenerator mitigates these issues in a couple of ways:
- It develops models of multiple "orders", that is, of multiple lengths of character sequences. If the generator encounters a new sequence of three characters like "jav", it will first check if it has trained a model on that sequence. If not, it will fall back to check if it has a model for "av", failing that, it will certainly have a model for what comes after "v". I call this a 3rd-order model and it is the default.
- A Bayesian prior probability is added to every character in the alphabet in every model, so some truly random character sequences not seen in the training data are possible. The alphabet is inferred from the training data, so any UTF-8 characters should be possible. The default prior is a relative probability of 0.005. Truly random output becomes more likely with a larger alphabet and with fewer trained character sequences, so you may want to play with this parameter: increase it to increase the randomness, or decrease it to make the output more like the training data.
MarkovTextGenerator ignores case, converting your input text and filters to lowercase and returning lowercase strings.
#### MarkovTextCasePreservingGenerator
A subclass of MarkovTextGenerator that learns and reproduces upper/lower case usage in the training data. With a given dataset, this model may learn less effectively from the training data because it builds separate models for "A" and "a" (to give an example) instead of combining observations. However, it may be preferable if the input data has interesting uses of capitalization (such as names that begin with "Mc" and "Mac" followed by capitals) that you want to re-generate. Any start/end filter(s) you configure will also be case-sensitive.
### ClusterChainGenerator
Quick start:
ClusterChainGenerator ccgen = new ClusterChainGenerator().train(myTextStream);
System.out.println(ccgen.generateOne());
(or with all the optional configuration...)
ClusterChainGenerator ccgen = new ClusterChainGenerator().withMaxOrder(2).withStartFilter("J").withEndFilter("ia").withMinLength(3).withMaxLength(15).withRandom(myRandom).train(myTextStream).andAddPriors(0.01);
System.out.println(ccgen.generateOne());
A class that uses a vowel/consonant clustering algorithm to generate new random text. Based loosely on [an algorithm described by Kusigrosz at RogueBasin](http://roguebasin.roguelikedevelopment.org/index.php?title=Cluster_chaining_name_generator), it scans input text for clusters of vowels and clusters of consonants, after converting it all to lowercase, keeping track of all clusters that have been observed to follow any given cluster. For example, "Elizabeth" would yield clusters `#-e-l-i-z-a-b-e-th-#` and "Anne" would yield `#-a-nn-e-#` where "`#`" is a control character marking the start or end of a string.
Much like MarkovTextGenerator, the implementation is based on a multi-order Markov chain (one difference is that priors aren't added by default and must be added explicitly after training). Internally we would keep track of the possible successors of each cluster, e.g.:
```
# -> [e,a]
e -> [l,th,#]
a -> [b,nn]
th -> [#]
...etc...
```
The `generateOne()` method takes a random walk through the cluster chain, only following paths that were found in the training data. To continue our example, a new string could begin with "e" or "a", with equal likelihood, an "e" could be followed by "l", by "th", or by the end of a string, and so on. With this training dataset of only two words, you could get a few different results, e.g.:
```
elizanneth
abelizanne
anneth
...etc...
```
Each newly generated candidate string is compared to filters (minLength, maxLength, startsWith, endsWith) and returned if it passes. If the candidate string is filtered out, we generate another, until one passes. (Be aware that if you configure very difficult-to-match filters, generation time may increase greatly. If you set up impossible-to-match filters, e.g. requiring characters that aren't in the training data set's alphabet, you will get an infinite loop.
### RandomDrawGenerator
Quick start:
RandomTextGenerator randomdraw = new RandomDrawGenerator().train(myTextStream);
System.out.println(randomdraw.generateOne());
(or with all the optional configuration...)
RandomTextGenerator randomdraw = new RandomDrawGenerator().withStartFilter("J").withEndFilter("ia").withMinLength(3).withMaxLength(15).withRandom(myRandom).train(myTextStream);
System.out.println(randomdraw.generateOne());
This generator simply draws a String at random from a `Stream<String>` of data fed into it. Useful, if not very sophisticated. Like MarkovTextGenerator, it allows the consumer to specify a desired minimum length, maximum length, start string, or end string, to filter the randomly-drawn text.
RandomDrawGenerator ignores case, converting your input text and filters to lowercase and returning lowercase strings.
### DoubleTextGenerator
Quick start:
RandomTextGenerator doubletext = new DoubleTextGenerator(
new MarkovTextGenerator().train(myFirstTextStream), // for example
new RandomDrawGenerator().train(myOtherTextStream), // for example
"-"
);
System.out.println(doubletext.generateOne());
This generator combines the output of two other RandomTextGenerators, which could be useful if you want to generate a combination of first name and last name, or a hyphenated name. Its constructor takes two RandomDrawGenerators and a String separator (if null, a single space is used by default).
## How to contribute
This package uses what I believe is the standard Maven file structure. If you fork and clone the repo, your IDE should be able to locate the `pom.xml` and the source and test files. To use maven to build and test it, simply
mvn clean test
If you'd like to build the JARs and Javadocs, and install the repo to your computer's Maven repository, run
mvn clean install
To contribute new code, corrections, etc., go ahead and make a pull request. New procedural generation algorithms would be welcomed (code them them as new classes implementing the RandomTextGenerator interface) as would new tests. If you have a great dataset of training data, please contribute that to the [examples](https://github.com/joeclark-phd/procedural-generation-examples) repository.
## Examples
I've built another project to run examples of output. You can find it here: [joeclark-net/procedural-generation-examples](https://github.com/joeclark-phd/procedural-generation-examples)
### MarkovTextGenerator examples
With **MarkovTextGenerator** trained on a file of 1360 ancient Roman names (/src/test/resources/romans.txt), order 3, prior 0.005F, minLength 4, maxLength 12, I generated these 25 names in 181ms (including the training):
caelis domidus pilianus naso recunobaro
potiti cerius petrentius herenialio caelius
venatius octovergilio favenaeus surus wasyllvianus
nentius soceanus lucia eulo atric
caranoratus melus sily fulcherialio dula
Setting the endFilter parameter to "a" filters out some passably female-sounding names. ("ia","na", and "la" are also good filters):
thea supera variwawrzma vediskozma isarina
tertia lasca juba lucia critula
nigelasca vagnenna armina salatera pulcita
cellasca verula ocessanga cimylla galla
mercuribosma limeta juba pulcita esdranicola
An alternative strategy is simply to train the generator on a single-sex dataset. Here on the left, for example, are the results of training the generator with a file of 146 female Viking names, and on the right, a generator trained on 498 male Viking names. (These training data can be found in [joeclark-net/procedural-generation-examples](https://github.com/joeclark-phd/procedural-generation-examples).)
FEMALE: MALE:
øviyrsa thorhild iorn sigfast
holm hallgeot hersi osvid
drid sibergljot øpir vald
halla drid solmsteinund hæmingjal
tonna sæurijorgärd sumävf boel
freyngtdrun kiti slärdar hundi
grelod ingulfrid kjxim kætilbisld
asvid hard soälverkvott sumarlid
fastrid gudland sigfus kveldun
hild inga torsteinth wary
geirhild ginna spjut sjägfiæmund
ingeltorg ingrta hromund orleif
Note that the **MarkovTextGenerator** automatically infers an alphabet from the training data, including Scandinavian characters that aren't on my keyboard.
### ClusterChainGenerator examples
The ClusterChainGenerator does what the MarkovTextGenerator does, but with clusters of consonants or vowels instead of individual characters. Using the same training dataset of Roman names, I generated these 25 random Roman names in 83ms:
festus minus frumerinus clodius aebuteo
acitalina docilusius marcellus rectus placilius
tertulus brictius viber nepius salvian
burrus stertinus gordianus sevtonius protus
allobrogicus didicus christianus quietus hosidonax
## Release notes
**Release 1.1**
- added `ClusterChainGenerator`
**Release 1.0.1**
- `MarkovNameGenerator.train()` no longer clears the alphabet and model before running. Therefore, you can now train a model on multiple input streams.
**Release 1.0**
- initial launch including `MarkovTextGenerator`, `RandomDrawGenerator`, and `DoubleTextGenerator`
|
2024-11-18T18:20:10.719795+00:00 | 2022-05-27T16:09:39 | cec93b5e214e0d70124c016d48c99a9e96257ed0 | 4 | {
"blob_id": "cec93b5e214e0d70124c016d48c99a9e96257ed0",
"branch_name": "refs/heads/master",
"committer_date": "2022-05-27T16:09:39",
"content_id": "31d837a69fc9b4bb7803d557cbebe2bd671d99ef",
"detected_licenses": [
"MIT"
],
"directory_id": "669dc2374e8ac04a17c75caa0a310a05a4f7e13f",
"extension": "md",
"filename": "README.md",
"fork_events_count": 1,
"gha_created_at": "2017-11-06T15:54:40",
"gha_event_created_at": "2020-09-30T10:59:14",
"gha_language": "Shell",
"gha_license_id": "MIT",
"github_id": 109716448,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 2316,
"license": "MIT",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:16",
"repo_name": "itsankoff/ansible-role-letsencrypt",
"revision_date": "2022-05-27T16:09:39",
"revision_id": "fbef9b0c868ada7808979af8bfc25f1fe579c3de",
"snapshot_id": "9a8489c8c64f9126a84306acb36fb4c170b038fc",
"src_encoding": "UTF-8",
"star_events_count": 1,
"url": "https://raw.githubusercontent.com/itsankoff/ansible-role-letsencrypt/fbef9b0c868ada7808979af8bfc25f1fe579c3de/README.md",
"visit_date": "2022-06-16T22:33:31.088099"
} | 3.8125 | stackv2 | ansible-role-letsencrypt
=========
Easy to use letsencrypt ansible role to create and renew SSL certificates.
Current version is tested and works on __Ubuntu 16.04__ to __Ubuntu 20.04__.
In future more platforms will be added.
Requirements
------------
If you use the default `letsencrypt_mode` for creating ssl certificates then you __MUST__ enable http path `/.well-known` in your vhost configuration to allow letsencrypt to verify the ownership of your domains. If you use `standalone` mode then you don't need any changes in your vhost configuration. For more information about modes see __Role Variables__ section.
* Nginx example:
```
location ~ /.well-known {
allow all;
}
```
Role Variables
--------------
* __[REQUIRED]__ `{{ letsencrypt_domains }}` - list of domains for which the certificate will be obtained.
* __[REQUIRED]__ `{{ letsencrypt_email }}` - an email for receiving important announcements and notices.
* __[REQUIRED]__ `{{ letsencrypt_renewal_cron_dest }}` - the destination for renewal cron which renews all installed letsencrypt certificates.
* `{{ letsencrypt_mode }}` - the mode for issuing ssl certificates. Default mode is `webroot` which does not require to stop and start http service during ssl certificate creation. If you don't have any http services in you deployment you can switch to `standalone` mode.
* `{{ letsencrypt_production }}` - if `true`, letsencrypt creates live certificate ready to use. Otherwise it creates staging certificate.
* `{{ letsencrypt_http_service }}` - the http service which will use the certificates. The default value is `nginx`.
* `{{ letsencrypt_webroot_dir }}` - if you use `webroot` mode then you need to provide webroot directory to enable letsencrypt to verify domains ownership.
Dependencies
------------
NA
Example Playbook
----------------
```
- hosts: servers
roles:
- role: itsankoff.letsencrypt
letsencrypt_domains: ["example.com", "www.example.com"]
letsencrypt_email<EMAIL_ADDRESS> letsencrypt_renewal_cron_dest: "/opt"
```
License
-------
[MIT](https://github.com/itsankoff/ansible-role-letsencrypt/blob/master/LICENSE)
Author Information
------------------
For more information please contact me:
* Email<EMAIL_ADDRESS>* [Github](https://github.com/itsankoff)
|
2024-11-18T18:20:10.801188+00:00 | 2021-03-13T22:51:35 | 831b237a0576912e6ec5dc2291530d69337d5bfd | 3 | {
"blob_id": "831b237a0576912e6ec5dc2291530d69337d5bfd",
"branch_name": "refs/heads/main",
"committer_date": "2021-03-13T22:51:35",
"content_id": "c00fd5c58fcbd0d848c17b0a1bc23fb9ec18e1f3",
"detected_licenses": [
"MIT"
],
"directory_id": "4922ca1673343714d3401598554b9c6094e3b3b0",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": 347249999,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 1864,
"license": "MIT",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:17",
"repo_name": "thereal-baitjet/Work-Day-scheduler",
"revision_date": "2021-03-13T22:51:35",
"revision_id": "473503065efa49642f1b8a53c254af5e5990e32b",
"snapshot_id": "6e273cd3c25b8d3cd6de05fd66b5fe0461d4f9ff",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/thereal-baitjet/Work-Day-scheduler/473503065efa49642f1b8a53c254af5e5990e32b/README.md",
"visit_date": "2023-03-24T06:06:19.301676"
} | 2.515625 | stackv2 | # Work-Day-scheduler
## Description
I Created a functional day planner where you can input your tasks and keep track of time in order to work efficientlly.
when you enter a task it saves to your local storage. its color cotated to past present and future as a reminder to work diligently. also displays todays date.
## Instalation
enter this url https://thereal-baitjet.github.io/Work-Day-scheduler/

## credits
Ricardo Russ https://www.linkedin.com/in/ricardoljruss
Sylvia Prabudy https://github.com/sylviaprabudy
getbootstrap.com
w3schools.com
## License
MIT License
Copyright (c) [2021] [Juan Carlos Santos]
Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software"), to deal
in the Software without restriction, including without limitation the rights
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
copies of the Software, and to permit persons to whom the Software is
furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be included in all
copies or substantial portions of the Software.
IN NO EVENT SHALL THE
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
SOFTWARE.
## Badge

|
2024-11-18T18:20:10.848066+00:00 | 2022-07-18T17:29:49 | 746102383ecc55ddb80306a618d7800832cf69f3 | 4 | {
"blob_id": "746102383ecc55ddb80306a618d7800832cf69f3",
"branch_name": "refs/heads/master",
"committer_date": "2022-07-18T17:29:49",
"content_id": "2f850ff873aa8cb7a72c25d62cedd66925849ad8",
"detected_licenses": [
"MIT"
],
"directory_id": "943f0fd8070e930be83b7ba98a457f66c69f3f25",
"extension": "md",
"filename": "install.md",
"fork_events_count": 3,
"gha_created_at": "2018-07-11T16:03:47",
"gha_event_created_at": "2022-07-18T17:29:50",
"gha_language": "HTML",
"gha_license_id": "MIT",
"github_id": 140599158,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 1847,
"license": "MIT",
"license_type": "permissive",
"path": "/navigation/install.md",
"provenance": "stack-edu-markdown-0000.json.gz:18",
"repo_name": "mivaecommerce/Extensions",
"revision_date": "2022-07-18T17:29:49",
"revision_id": "04058709b414983df65cd7e8874d7214838879ee",
"snapshot_id": "7c36b17c4ab73b703e11c155ac8a06ae5885d736",
"src_encoding": "UTF-8",
"star_events_count": 1,
"url": "https://raw.githubusercontent.com/mivaecommerce/Extensions/04058709b414983df65cd7e8874d7214838879ee/navigation/install.md",
"visit_date": "2022-11-10T04:19:23.304623"
} | 3.640625 | stackv2 | # Transfigure Navigation Installation
If your ReadyTheme did not come with the default Transfigure Navigation extension pre-installed, you can add it to your site by performing the following steps.
### Update Transfigure Navigation HTML
- Navigate to User Interface -> Theme Components -> Navigation Sets -> primary_navigation
- Paste the content from [transfigure-navigation.mvt](transfigure-navigation.mvt).
### Add Transfigure Navigation to global header
- Navigate to User Interface -> Global Settings -> Settings -> Global Header
- Paste the following code where you would like the navigation to appear:
- _Note: You may need to modify to fit best in your theme._
```html
<div class="o-wrapper o-wrapper--full o-wrapper--flush t-site-navigation &mvte:global:checkout_hidden;">
<div class="o-layout o-layout--align-center o-layout--flush o-layout--justify-center t-site-navigation__wrap">
<mvt:item name="readytheme" param="navigationset( 'primary_navigation' )" />
</div>
</div>
<!-- end t-site-navigation -->
```
### Add the CSS
- Create a new build subdirectory as `build/extensions/transfigure-navigation/`
- Add [_transfigure-navigation.scss](_transfigure-navigation.scss) to the directory.
- Update your `build/extensions/extensions.scss` file to include `@import "navigation/transfigure-navigation";`
### Add the JavaScript
- Add [transfigure-navigation.js](transfigure-navigation.js) to your `build/extensions/navigation/` subdirectory.
- Update your `gulpfile.js` file to include `buildPathExtensions + '/navigation/transfigure-navigation.js'` to your `let extensions` variable.
### Initialize the extension
Within `build/ui/theme.js`, locate `init: function () {` and append this code:
```javascript
/**
* Initialize the Transfigure Navigation extension
*/
$.hook('has-drop-down').transfigureNavigation();
```
|
2024-11-18T18:20:10.973762+00:00 | 2018-12-17T16:35:05 | 5c7f59f21a23e6cc844c6d812c721daf78785051 | 4 | {
"blob_id": "5c7f59f21a23e6cc844c6d812c721daf78785051",
"branch_name": "refs/heads/master",
"committer_date": "2018-12-17T16:35:05",
"content_id": "b6a8031ac0778135c8e3214edbcc911dc5d57ee8",
"detected_licenses": [
"MIT"
],
"directory_id": "52fe5485b2b535f362e4cb4e7c8fed4f347e9e20",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": 139327839,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 2754,
"license": "MIT",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:19",
"repo_name": "roosmaa/bitcharge-rs",
"revision_date": "2018-12-17T16:35:05",
"revision_id": "bb349107ffccf2f613e141a5d7dc3652a7841242",
"snapshot_id": "649a8fcda54d39462ece47f30ceb9789a0324520",
"src_encoding": "UTF-8",
"star_events_count": 3,
"url": "https://raw.githubusercontent.com/roosmaa/bitcharge-rs/bb349107ffccf2f613e141a5d7dc3652a7841242/README.md",
"visit_date": "2020-03-22T01:45:33.976401"
} | 3.78125 | stackv2 | # BitCharge - Payment page for accepting bitcoins
BitCharge is a small web app that allows your clients to pay you in bitcoins, which are then automatically converted to euros and deposited to your bank account. The conversion is handled by the [Coinmotion](https://coinmotion.com/) exchange.

This software is useful for freelancers and small companies:
- ... who invoice their clients manually
- ... who reside in the eurozone
- ... who are tech savvy (managing a virtual server, comfortable with command-line and ssh)
BitCharge is an alternative to BitPay, CoinPayments and other SaaS offerings. Those services are more tailored towards the retail sector. For a small company it doesn't make sense to pay their fees nor the extra fees of the bitcoin network as money is moved around.
It is still very early days of BitCharge, but it currently has the following features:
- Clean payment instructions page for the payer
- Quoting a bitcoin amount based on the live EUR/BTC exchange rates
- Funds are paid directly to your Coinmotion account, avoiding unnecessary fees
- Automatically covert bitcoins to euros and withdraw them immediately via the Coinmotion API
## Setup instructions
- Make sure you have Rust programming language installed ([instructions](https://www.rust-lang.org/en-US/install.html))
- Clone this repository
- Run `cargo build --release`
- Copy `bitcharge.toml.example` to `bitcharge.toml`
- Use your preferred text editor to edit `bitcharge.toml` and fill out the required fields
- Start the app `RUST_LOG=info ./target/release/bitcharge`
You should also setup Nginx/Apache with HTTPS termination in front of BitCharge.
### Registering invoices with BitCharge
Registering invoices with BitCharge is somewhat tedious currently, but it will get better in the future.
- Open Coinmotion web interface, go to Receive tab and generate a new deposit address using your invoice number as the description
- Open the `bitcharge.toml` in your preferred text editor and add a new `[[charges]]` section to it (template below), filling it in with the appropriate information and your newly generated deposit address
- Restart the _bitcharge_ process and monitor the logs for the list of charges to get the public URL associated with your new charge
Template for the new charge to be used in the `bitcharge.toml` file:
```
[[charges]]
# The ID should be incremented for each new charge (1, 2, 3, ...)
id = 1
# The unique identifier on your invoices
invoice_id = "2018-0012"
# The amount in euros that the invoice is made out to be
eur_amount = "1234.56"
# The new deposit address from Coinmotion
btc_address = ""
```
|
2024-11-18T18:20:11.020554+00:00 | 2021-01-06T08:44:04 | 966cac8b3d13081466db4918aacf5decdda9052b | 3 | {
"blob_id": "966cac8b3d13081466db4918aacf5decdda9052b",
"branch_name": "refs/heads/2.x",
"committer_date": "2021-01-06T09:05:19",
"content_id": "ec568c51d3e42e00280cef1d14406be64e53ff30",
"detected_licenses": [
"Apache-2.0"
],
"directory_id": "cf7f3e3790484a437edd0db5a883f375692d65e4",
"extension": "md",
"filename": "README.md",
"fork_events_count": 11,
"gha_created_at": "2013-06-02T05:13:54",
"gha_event_created_at": "2021-01-06T09:12:30",
"gha_language": "PHP",
"gha_license_id": "Apache-2.0",
"github_id": 10432812,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 2698,
"license": "Apache-2.0",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:20",
"repo_name": "skymeyer/Vatsimphp",
"revision_date": "2021-01-06T08:44:04",
"revision_id": "60eb40c0bcf7dd58060ae529073fa25eb455dedd",
"snapshot_id": "7ee39ad9b8c5cf164b54347baf28bbfc630a4514",
"src_encoding": "UTF-8",
"star_events_count": 27,
"url": "https://raw.githubusercontent.com/skymeyer/Vatsimphp/60eb40c0bcf7dd58060ae529073fa25eb455dedd/README.md",
"visit_date": "2021-01-22T01:05:16.227442"
} | 3.359375 | stackv2 | # Vatsimphp [](https://packagist.org/packages/skymeyer/vatsimphp) [](https://packagist.org/packages/skymeyer/vatsimphp) [](https://travis-ci.com/skymeyer/Vatsimphp)
Vatsimphp collects and parses the publically available statistics
from the [VATSIM.net](http://www.vatsim.net) network. It provides
iterators for the available data and basic search/filtering
capabilities without using a database backend.
Vatsimphp uses an intelligent local file cache to avoid consuming
unnecessary bandwidth from the public data servers. The software
comes with sensible default settings, but can be changed if
required. Vatsimphp can be dropped into your crontab very easily
to avoid inline data updates in your web application.
A "cache only" node is available if another process
is already responsible to retrieve the raw data files from the
VATSIM network or if live connections are not applicable.
Vatsimphp can be easily plugged into existing PHP systems to
query VATSIM data or feed the parsed results into a database
backend of your choice.
Documentation
-------------
- [Browse documentation](https://github.com/skymeyer/Vatsimphp/blob/2.x/docs/index.md)
- [Browse examples](https://github.com/skymeyer/Vatsimphp/tree/2.x/examples)
About
=====
Requirements
------------
- From PHP 7.2 and above
- PHPUnit 7.4 or higher for test suite execution
Bugs and feature requests
-------------------------
Bugs and feature request can be filed on the [issues](https://github.com/skymeyer/Vatsimphp/issues) page.
Contributing
------------
Contributions are welcome in respect of the [PSR-2](https://github.com/php-fig/fig-standards/blob/master/accepted/PSR-2-coding-style-guide.md)
coding style and unit test coverage.
Changelog
---------
See the `CHANGELOG.md` file for more details.
Author
------
Jelle Vink -<EMAIL_ADDRESS>(<http://skymeyer.dev>)
License
-------
Vatsimphp is licensed under the Apache License, Version 2.0. Check the `LICENSE` and `NOTICE` file for details.
Disclaimer
----------
VATSIM (Virtual Air Traffic Simulation Network) is a non-profit organisation.
The development of this software is not directly affiliated to the VATSIM
organisation and provided on an "as is" basis as set forward in the above License.
This software does not connect to the VATSIM network directly, but rather consumes
publically available resources produced by the VATSIM network.
More information on VATSIM can be found at <http://www.vatsim.net>.
|
2024-11-18T18:20:11.061830+00:00 | 2017-10-24T14:10:46 | f46d245bdf5cd3ab93e9dc2e4933f5ccd92f0242 | 3 | {
"blob_id": "f46d245bdf5cd3ab93e9dc2e4933f5ccd92f0242",
"branch_name": "refs/heads/master",
"committer_date": "2017-10-24T14:10:46",
"content_id": "2ac925527df68ea4170a00df11b7037a6eac6a1c",
"detected_licenses": [
"MIT"
],
"directory_id": "70417b31741ae87244602d04892ffea168c1ed8d",
"extension": "md",
"filename": "ROADMAP.md",
"fork_events_count": 1,
"gha_created_at": "2017-09-02T07:05:09",
"gha_event_created_at": "2022-08-05T12:32:07",
"gha_language": "Go",
"gha_license_id": "MIT",
"github_id": 102179359,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 3919,
"license": "MIT",
"license_type": "permissive",
"path": "/vendor/github.com/alexellis/faas/ROADMAP.md",
"provenance": "stack-edu-markdown-0000.json.gz:21",
"repo_name": "kenfdev/faas-rancher",
"revision_date": "2017-10-24T14:10:46",
"revision_id": "29aae48a100e52e0c1f4c5627cf597f15b33c00b",
"snapshot_id": "eb00039127e831ac44f85189068446ac0b8a6436",
"src_encoding": "UTF-8",
"star_events_count": 32,
"url": "https://raw.githubusercontent.com/kenfdev/faas-rancher/29aae48a100e52e0c1f4c5627cf597f15b33c00b/vendor/github.com/alexellis/faas/ROADMAP.md",
"visit_date": "2022-08-26T05:38:15.964572"
} | 3.375 | stackv2 | # Roadmap
## 1. Current features
For an overview of features in August 2017 read the following post:
* [Introducing Functions as a Service (FaaS)](https://blog.alexellis.io/introducing-functions-as-a-service/)
## GitHub repos:
* https://github.com/alexellis/faas
* https://github.com/alexellis/faas-netes
* https://github.com/alexellis/faas-cli
* https://github.com/openfaas/nats-queue-worker
### The API Gateway
This container acts in a similar way to the API Gateway on AWS. Requests can be made to this endpoint with a JSON body.
Some of the more recent Completed Features:
* UI for viewing and testing functions deployed through stack
* Auto-scaling of replicas as load increases
* Backing off of replicas as load reduces
* Unique URL routes for serverless functions
* Instrumentation via Prometheus metrics at GET /metrics
* Bundled Prometheus stack with AlertManager
* UI enhancements to create new function through a form
* Raspberry Pi (armhf/armv6) support (currently available)
* Documentation for current API in Swagger format
* Documentation about [Alexa sample function](https://blog.alexellis.io/serverless-alexa-skill-mobymingle/)
* Native CLI for templating/building and deploying functions
* Windows support for watchdog back-end - so that Windows executables can be used in a multi-OS swarm
* Enforcing function execution time in seconds.
* Python, Node.js, Ruby and CSharp code templates for the CLI
* Delete function in CLI
* Developer guide for CSharp
* Developer guide for Python
* Kubernetes support
**Incoming requests and routing**
There are three options for routing:
* Functions created on the overlay network can be invoked by: http://localhost:8080/function/{servicename}
* Automatic routing is also enabled through the `/` endpoint via a `X-Function` header which matches a service name (function) directly.
### The watchdog
This binary fwatchdog acts as a watchdog for your function. Features:
* Static binary in Go
* Listens to HTTP requests over swarm overlay network
* Spawns process set in `fprocess` ENV variable for each HTTP connection
* Only lets processes run for set duration i.e. 500ms, 2s, 3s.
* Language/binding independent - can invoke any UNIX process, including built-ins such as `wc` or `cat`
## 2. Future items
Most items are detailed [via Github issues](https://github.com/alexellis/faas/issues).
Must have
* Re-branding to OpenFaaS
* New logo - graphic icon and text (in progress)
* Website / landing page (in progress)
* Asynchronous / long-running tasks (PR in testing)
Should have
* AARCH64 (64-bit ARM) port (dependent on Docker release schedule)
* Integration with a reverse proxy - such as Traefik or Kong
* Basic auth for /system endpoints (probably via reverse proxy)
* CLI - list functions / query function info
* OS constraints in the deploy function API
* Healthchecks for functions deployed on Kubernetes
Could have
* Built-in Docker registry with default configuration
* Docker image builder (remote service)
* Function store - list of useful predefined functions
* Supporting request parameters
* Configurable memory limits via "new function" pop-up (already supported by Docker compose stack)
Nice to have
* Developer Cloud guide:
* for Digital Ocean
* for Packet
* Developer guide for your first Node.js function
* Developer guide to using functions together - via pipes on client, or a "director" function on server
* Documentation on using CRON / JenkinsCI for invoking functions on a timed basis
* Guide for termination through NGinx or built-in TLS termination
* Guide for basic authentication over HTTPs (set up externally through NGinx etc)
* CRIU - (Checkpoint/Restore In Userspace) for warm-loading serverless tasks with a high start-up cost/latency.
* Deeper tests coverage and integration tests
### Contributing
Please see [CONTRIBUTING.md](https://github.com/alexellis/faas/blob/master/CONTRIBUTING.md).
|
2024-11-18T18:20:11.199407+00:00 | 2021-06-15T15:45:26 | 2e5c4c45d7184bdb2c0a60fd9ee7210bd501a529 | 4 | {
"blob_id": "2e5c4c45d7184bdb2c0a60fd9ee7210bd501a529",
"branch_name": "refs/heads/main",
"committer_date": "2021-06-15T15:45:26",
"content_id": "f0580d55bc130ab6c2da6196e9fa92c3d87040ef",
"detected_licenses": [
"MIT"
],
"directory_id": "85f9c60c1c04b865bf9ae872516c49eea0130849",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": 370077192,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 4265,
"license": "MIT",
"license_type": "permissive",
"path": "/Hyewon_seg/DPT/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:22",
"repo_name": "soohi0/ComputerVision1team",
"revision_date": "2021-06-15T15:45:26",
"revision_id": "173a729106df2d1b33ac673dfa7f4c2c7a199e4c",
"snapshot_id": "6af5ea5f9ff20dfd98086dd5693b33d53207ac1f",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/soohi0/ComputerVision1team/173a729106df2d1b33ac673dfa7f4c2c7a199e4c/Hyewon_seg/DPT/README.md",
"visit_date": "2023-06-04T08:15:22.427440"
} | 3.578125 | stackv2 | ## Vision Transformers for Dense Prediction
This repository contains code and models for our [paper](https://arxiv.org/abs/2103.13413):
> Vision Transformers for Dense Prediction
> René Ranftl, Alexey Bochkovskiy, Vladlen Koltun
### Changelog
* [March 2021] Initial release of inference code and models
### Setup
0) I recommend you make virtual environment. and activate it.
- $ conda create -n 가상환경이름 python=버전(python version must be 3.6 or 3.7)
- $ conda activate 가상환경 이름
- 만약에 가상환경 list를 확인하고 싶을 때 -> conda info -e
- 가상환경 종료 시 : conda deactivate
1) when you create virtual environment, install conda, opencv
- conda install conda
- conda install openCV
2) Download the model weights and place them in the `weights` folder:
Monodepth:
- [dpt_hybrid-midas-501f0c75.pt](https://github.com/intel-isl/DPT/releases/download/1_0/dpt_hybrid-midas-501f0c75.pt), [Mirror](https://drive.google.com/file/d/1dgcJEYYw1F8qirXhZxgNK8dWWz_8gZBD/view?usp=sharing)
- [dpt_large-midas-2f21e586.pt](https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-midas-2f21e586.pt), [Mirror](https://drive.google.com/file/d/1vnuhoMc6caF-buQQ4hK0CeiMk9SjwB-G/view?usp=sharing)
Segmentation:
- [dpt_hybrid-ade20k-53898607.pt](https://github.com/intel-isl/DPT/releases/download/1_0/dpt_hybrid-ade20k-53898607.pt), [Mirror](https://drive.google.com/file/d/1zKIAMbltJ3kpGLMh6wjsq65_k5XQ7_9m/view?usp=sharing)
- [dpt_large-ade20k-b12dca68.pt](https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-ade20k-b12dca68.pt), [Mirror](https://drive.google.com/file/d/1foDpUM7CdS8Zl6GPdkrJaAOjskb7hHe-/view?usp=sharing)
3) Set up dependencies:
(방법 1)
```shell
conda install pytorch torchvision opencv
```
(actually this shell is for GPU environment, if your environment is CPU and you dont have CUDA, then shell is down.
:참고 : https://pytorch.org/get-started/locally/)
(방법 2)
```shell
$ conda install pytorch torchvision torchaudio cpuonly -c pytorch
```
```shell
pip install timm
```
The code was tested with Python 3.7, PyTorch 1.8.0, OpenCV 4.5.1, and timm 0.4.5
in Hyewon Version_ Python 3.6, Pytorch 1.7.1, OpenCV 3.3.1, and timm 0.4.9
### Usage
1) Place one or more input images in the folder `input`.
- in this version, dir 'input_imgs' folder have images.
- So copy those imgs to folder 'input'.
- if you want to test more images, then input more images input.
2) Run a monocular depth estimation model:
```shell
python run_monodepth.py
```
Or run a semantic segmentation model:
```shell
python run_segmentation.py
```
3) The results are written to the folder `output_monodepth` and `output_semseg`, respectively.
Use the flag `-t` to switch between different models. Possible options are `dpt_hybrid` (default) and `dpt_large`.
**Additional models:**
- Monodepth finetuned on KITTI: [dpt_hybrid_kitti-cb926ef4.pt](https://github.com/intel-isl/DPT/releases/download/1_0/dpt_hybrid_kitti-cb926ef4.pt)
- Monodepth finetuned on NYUv2: [dpt_hybrid_nyu-2ce69ec7.pt](https://github.com/intel-isl/DPT/releases/download/1_0/dpt_hybrid_nyu-2ce69ec7.pt)
Run with
```shell
python run_monodepth -t [dpt_hybrid_kitti|dpt_hybrid_nyu]
```
### Citation
Please cite our papers if you use this code or any of the models.
```
@article{Ranftl2021,
author = {Ren\'{e} Ranftl and Alexey Bochkovskiy and Vladlen Koltun},
title = {Vision Transformers for Dense Prediction},
journal = {ArXiv preprint},
year = {2021},
}
```
```
@article{Ranftl2020,
author = {Ren\'{e} Ranftl and Katrin Lasinger and David Hafner and Konrad Schindler and Vladlen Koltun},
title = {Towards Robust Monocular Depth Estimation: Mixing Datasets for Zero-shot Cross-dataset Transfer},
journal = {IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI)},
year = {2020},
}
```
### Acknowledgements
Our work builds on and uses code from [timm](https://github.com/rwightman/pytorch-image-models) and [PyTorch-Encoding](https://github.com/zhanghang1989/PyTorch-Encoding). We'd like to thank the authors for making these libraries available.
### License
MIT License
|
2024-11-18T18:20:11.229376+00:00 | 2022-02-20T23:03:41 | da1aa0dc18b055d049519a18588b0d7965bbf4bb | 3 | {
"blob_id": "da1aa0dc18b055d049519a18588b0d7965bbf4bb",
"branch_name": "refs/heads/master",
"committer_date": "2022-02-20T23:03:41",
"content_id": "4e38357fdd6e9c5bd6989abe603dc7210cfae54c",
"detected_licenses": [
"MIT"
],
"directory_id": "4da3eca681d465f4994e84f43472cf66664e9b02",
"extension": "md",
"filename": "a599c8752b4dd815b68c9ba54667297009e67c2755c2f0e86efc959dcd6064e1.md",
"fork_events_count": 2,
"gha_created_at": "2018-06-20T20:02:33",
"gha_event_created_at": "2023-04-12T05:47:06",
"gha_language": null,
"gha_license_id": "MIT",
"github_id": 138080116,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 1663,
"license": "MIT",
"license_type": "permissive",
"path": "/_emails/1995/12/a599c8752b4dd815b68c9ba54667297009e67c2755c2f0e86efc959dcd6064e1.md",
"provenance": "stack-edu-markdown-0000.json.gz:23",
"repo_name": "cryptoanarchywiki/mailing-list-archive-generator",
"revision_date": "2022-02-20T23:03:41",
"revision_id": "5ee11c76b130aadf0ed74877107df5053ab0361b",
"snapshot_id": "7547cba9dd353f2f6dcc64eab77473ed5967c734",
"src_encoding": "UTF-8",
"star_events_count": 5,
"url": "https://raw.githubusercontent.com/cryptoanarchywiki/mailing-list-archive-generator/5ee11c76b130aadf0ed74877107df5053ab0361b/_emails/1995/12/a599c8752b4dd815b68c9ba54667297009e67c2755c2f0e86efc959dcd6064e1.md",
"visit_date": "2023-04-16T03:14:42.652593"
} | 2.578125 | stackv2 | ---
layout: default
---
# 1995-12-19 - Please help
## Header Data
From: Rich Salz<EMAIL_ADDRESS>To<EMAIL_ADDRESS>Message Hash: a599c8752b4dd815b68c9ba54667297009e67c2755c2f0e86efc959dcd6064e1<br>
Message ID<EMAIL_ADDRESS>Reply To: _N/A_<br>
UTC Datetime: 1995-12-19 23:05:18 UTC<br>
Raw Date: Wed, 20 Dec 1995 07:05:18 +0800<br>
## Raw message
```
{% raw %}From: Rich Salz<EMAIL_ADDRESS>Date: Wed, 20 Dec 1995 07:05:18 +0800
To<EMAIL_ADDRESS>Subject: Please help
Message-ID<EMAIL_ADDRESS>MIME-Version: 1.0
Content-Type: text/plain
Please help. Somehow, someone signed up the "cypherpunks" mailing
list to your "ZD Net Update" electronic newsletter. Our mailing list,
with over a thousand members, is for the discussion of cryptography
and technical means of achieving privacy and anonymity in the
emergent digital world.
There is some overlap between your publication and our mailing list,
which is why some well-meaning individual "signed us up." It's pretty
slight, however, and interested parties can sign up individually.
So, please remove<EMAIL_ADDRESS>from your mailing list. As
always, if you ahve questions about the propriety of this, you can
follow the internet conventions of writing to<EMAIL_ADDRESS>to double-check.
Thanks.
/r$
{% endraw %}
```
## Thread
+ Return to [December 1995](/archive/1995/12)
+ Return to "[Rich Salz <rsalz<span>@</span>osf.org>](/authors/rich_salz_rsalz_at_osf_org_)"
+ 1995-12-19 (Wed, 20 Dec 1995 07:05:18 +0800) - Please help - _Rich Salz<EMAIL_ADDRESS>
|
2024-11-18T18:20:11.265528+00:00 | 2023-06-09T16:55:18 | da8dea29eb4af99adcaa0185a4bd2a512b558008 | 3 | {
"blob_id": "da8dea29eb4af99adcaa0185a4bd2a512b558008",
"branch_name": "refs/heads/master",
"committer_date": "2023-06-09T16:55:18",
"content_id": "fb3bd48d10a3f064df4ec7bedc08210a69175f41",
"detected_licenses": [
"MIT"
],
"directory_id": "1c7b958545b7499dd24f107df444419b91f29958",
"extension": "markdown",
"filename": "2013-03-17-the-death-of-google-reader.markdown",
"fork_events_count": 0,
"gha_created_at": "2017-08-19T02:26:37",
"gha_event_created_at": "2022-10-27T21:40:32",
"gha_language": "HTML",
"gha_license_id": "MIT",
"github_id": 100764314,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 7669,
"license": "MIT",
"license_type": "permissive",
"path": "/_posts/2013-03-17-the-death-of-google-reader.markdown",
"provenance": "stack-edu-markdown-0000.json.gz:24",
"repo_name": "tonyduckles/blog_source",
"revision_date": "2023-06-09T16:55:18",
"revision_id": "e3777fd214bfcb78dc7fb0592d8afd2f1b40738a",
"snapshot_id": "798238544517ec5b82340fb2f856f7f7251ed357",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/tonyduckles/blog_source/e3777fd214bfcb78dc7fb0592d8afd2f1b40738a/_posts/2013-03-17-the-death-of-google-reader.markdown",
"visit_date": "2023-06-26T05:40:03.227068"
} | 2.90625 | stackv2 | ---
layout: post
title: "The Death of Google Reader"
date: 2013-03-17T13:57:22-05:00
comments: true
description: "Thoughts on the imminent closing of Google Reader, and the potential for new great alternatives."
categories:
- Web
---
So, Google is [killing Google Reader](http://googlereader.blogspot.com/2013/03/powering-down-google-reader.html).
When I saw that headline on Wednesday evening, I felt a mixture of emotions:
anger, sadness, and worry. Anger that Google could shut-down a service that
was so personally-valuable to me; sadness that a web-service I've depending
upon for years was going away; worry that there might not be an alternative
service out there that could fill the same role in all the same ways that the
Google Reader ecosystem has.
<!-- more -->
You see, I've been [using Google Reader since 2007](/blog/2007/05/19/google-reader/).
It's the web-app I use more than any others. I've had a dedicated Google
Reader pinned-tab open in all my Firefox sessions for as long as I can remember.
I use it every single day. It's the way I read the Internet.
Google Reader is more than just a simple web-app: it's the back-end RSS
aggregator service which crawls all the 120+ RSS feeds I follow and centrally
stores all the state about what unread (and starred) articles I have. That
kind of *read-anywhere-sync-to-everywhere* workflow is immensely valuable to me
because I want to be able to access my news wherever I am and with whatever
device is most handy: I use the Google Reader web-app on my home laptop, home
desktop, and work desktop machines; I use [Reeder](http://reederapp.com/) on my
iPhone; I use [gReader](https://play.google.com/store/apps/details?id=com.noinnion.android.greader.readerpro&hl=en)
on my Nexus 7 Android.
## Community Reactions
It's been interesting to read the reactions/backlash from the tech community. I
found some interesting blog-posts via Twitter:
### The Social Web Isn't a Replacement for RSS
Christian Heilmann wrote "[RIP Google Reader - I'd Have Paid For You](http://christianheilmann.com/2013/03/14/rip-google-reader-id-have-paid-for-you/)"
where we talks about how the social web will never be a replacement for the
one-stop aggregation service that Google Reader provides:
> Yes, RSS has been declared dead many times and people keep banging on about the social web and that Facebook, Twitter, Reddit and others have replaced the old style of blogging and having an own feed. But I don’t buy it, sorry. <b>Every social network is full of senseless chatter and organised advertising.</b> Social media experts and PR folk make sure that information about certain products and celebrities get read and retweeted. I don’t care about that. <b>I don’t want it. The same way I don’t watch public access channels or randomly surf channels but instead plan what I want to see on TV.</b> Random exploration and finding things by chance is fun, but it is not helping you to keep up to date – it is the ADHD of information consumption.
### Moving Towards a Walled-Off Web?
Jonathan Poritsky wrote a fantastic blog-post titled "[Reader's End and Google Today](http://www.candlerblog.com/2013/03/15/google-and-reader/)",
where he points out that this could be the start of a disturbing shift in
Google's priorities:
> The biggest issue doesn’t seem to be the loss of Reader itself, but the recognition <b>that Google’s priorities have shifted</b>
> ...
>
> But walls have sprouted up. Google can’t access the massive amounts of data people pour into Facebook and Twitter, so they built Google+ as their own social walled garden. Twitter is exerting control over how users experience their product, which shuts out competitors like Instagram (which is owned by Facebook), which can no longer display images inline in tweets. <b>The Web is getting smaller, not bigger, with each company working to become the umbrella under which you experience the Internet.</b> So Google has taken steps to make sure that the Web as users know it exists under their company banner, and Reader doesn’t fit in with that plan anymore.
>
> I was once a Google cheerleader. <b>Like many I believed their goal was to make a better Web for everyone</b>, with the one major tradeoff being that they would sell ads instead of charging users. That may once have been true but the Google of 2013 doesn’t want to build a better Web, it wants to build a better Google. I don’t think that goal is aligned with any of my own.
{% comment %}
Jonathan calls-out Anil Dash's blog-post titled "[The Web We Lost](http://dashes.com/anil/2012/12/the-web-we-lost.html)"
as a must-read article which I totally agree with. The times they are a-changin':
> The tech industry and its press have treated the rise of billion-scale social networks and ubiquitous smartphone apps as an unadulterated win for regular people, a triumph of usability and empowerment. They seldom talk about what we've lost along the way in this transition, and I find that younger folks may not even know how the web used to be.
{% endcomment %}
With this move, Google is seeding a lot of ill-will in the tech community.
This feels very much against the "Do No Evil" slogan which Google touts.
If Google can shutdown a service as beloved as Reader, then it makes you
wonder which services are safe from the chopping-block...
## A New Hope: Google Reader Alternatives
Even though Google says that "*usage of Google Reader has declined*", there's
obviously still a **lot** of people in the tech community who still find great
value in RSS and a Reader-like service. And a lot can happen in the next 3
months leading up to the July 1st shutdown.
So, the imminent death of Google Reader could just be trigger-point needed to
spark another renaissance in RSS readers like what we had circa 2005-2006.
And it looks like we will have several options. **[ReplaceReader](http://www.replacereader.com/)**
is a neat little site I found for folks to suggest replacements for Google Reader.
I expect the biggest challenge (for me personally) in finding a suitable
replacement for will be finding a solution that I can still (easily) access
(and seamlessly sync!) across multiple platforms. I've gotten spoiled-rotten
by the native iOS and Android Google Reader clients. And whatever I pick, I
want to make sure that I still have some exit-paths in case *that service*
closes-up shop for some reason. Some kind of open-source/self-host option could
be nice so that I can control my own data, but then again needing to maintain a
DB-backed website isn't really something I want to do anymore.
**[Tiny Tiny RSS](http://tt-rss.org/)** is an open-source self-host option which
looks pretty mature. This looks promising if all you need is a simple web interface.
The most interesting option I've seen so far is **[NewsBlur](http://www.newsblur.com)**.
It's open-source (on Github) and self-host-able (nice to know I have options)
but also has a paid hosted option. And it looks extremely polished and
**simply gorgeous**. It keeps the same simple/functional interface principals as
Google Reader while updating the UI for 2013. And there are iOS and Android
clients so that I can still access my news on whatever device I want.
I expect (or at least hope) there will be a flurry of activity in the RSS
reader space in the next few months leading up to the July 1st shutdown. It
will be interesting to see what alternatives the community embraces. I plan to
watch the space for a while before committing to any particular option, to see
which options rise to the top.
I plan to write a follow-up post in a few months detailing what option I end-up
going with...
|
2024-11-18T18:20:11.663837+00:00 | 2018-04-27T07:01:12 | 24b9ec11919e8a26e61c422bd49a51157e870259 | 2 | {
"blob_id": "24b9ec11919e8a26e61c422bd49a51157e870259",
"branch_name": "refs/heads/master",
"committer_date": "2018-04-27T07:01:12",
"content_id": "8cd899bd34fb49c5d70dfc171237dac863450342",
"detected_licenses": [
"Apache-2.0"
],
"directory_id": "190c4cfcf2b8a3425889b21b1411c69932f3867c",
"extension": "md",
"filename": "README.md",
"fork_events_count": 5,
"gha_created_at": "2017-12-24T01:28:56",
"gha_event_created_at": "2018-01-29T22:05:57",
"gha_language": "Smarty",
"gha_license_id": null,
"github_id": 115232471,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 316,
"license": "Apache-2.0",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:25",
"repo_name": "yassan/rancher-catalog",
"revision_date": "2018-04-27T07:01:12",
"revision_id": "960bc4ae91af184f4c2ded168de142164a7ee594",
"snapshot_id": "d3cd16935cceade20276de516179bd4ca2f8a8e5",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/yassan/rancher-catalog/960bc4ae91af184f4c2ded168de142164a7ee594/README.md",
"visit_date": "2021-09-13T08:38:36.346780"
} | 2.171875 | stackv2 | Rancher Yassan's Catalog
========================
This catalog provides templates created by yassan.
cf. [Rancher](https://github.com/rancher)
## templates
- [getredash/redash: Make Your Company Data Driven. Connect to any data source, easily visualize and share your data.](https://github.com/getredash/redash)
|
2024-11-18T18:20:11.892148+00:00 | 2019-09-05T02:06:29 | 347c960b97ac6573ce5d027d80fe8c204e5c606b | 3 | {
"blob_id": "347c960b97ac6573ce5d027d80fe8c204e5c606b",
"branch_name": "refs/heads/master",
"committer_date": "2019-09-05T02:06:29",
"content_id": "7e8321bab2d4e042704da132505cfa58e4a0ca98",
"detected_licenses": [
"MIT"
],
"directory_id": "eb6d91c4eb4d07edbd0e16addf0bf15c008e6e88",
"extension": "md",
"filename": "2019-04-19-Query-focused-Sentence-Compression-in-Linear-Time.md",
"fork_events_count": 4,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": 113123624,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 1083,
"license": "MIT",
"license_type": "permissive",
"path": "/_posts/2019-04-19-Query-focused-Sentence-Compression-in-Linear-Time.md",
"provenance": "stack-edu-markdown-0000.json.gz:26",
"repo_name": "AMDS123/amds123.github.io",
"revision_date": "2019-09-05T02:06:29",
"revision_id": "80ccfe8c852685e4829848229b22ba4736c65a7c",
"snapshot_id": "74e33be33d80aae2af58647dd33472de9e91310e",
"src_encoding": "UTF-8",
"star_events_count": 8,
"url": "https://raw.githubusercontent.com/AMDS123/amds123.github.io/80ccfe8c852685e4829848229b22ba4736c65a7c/_posts/2019-04-19-Query-focused-Sentence-Compression-in-Linear-Time.md",
"visit_date": "2021-10-18T22:26:01.031886"
} | 2.953125 | stackv2 | ---
layout: post
title: "Query-focused Sentence Compression in Linear Time"
date: 2019-04-19 02:19:43
categories: arXiv_CL
tags: arXiv_CL Face
author: Abram Handler, Brendan O'Connor
mathjax: true
---
* content
{:toc}
##### Abstract
Search applications often display shortened sentences which must contain certain query terms and must fit within the space constraints of a user interface. This work introduces a new transition-based sentence compression technique developed for such settings. Our method constructs length and lexically constrained compressions in linear time, by growing a subgraph in the dependency parse of a sentence. This approach achieves a 4x speed up over baseline ILP compression techniques, and better reconstructs gold shortenings under constraints. Such efficiency gains permit constrained compression of multiple sentences, without unreasonable lag.
##### Abstract (translated by Google)
##### URL
[http://arxiv.org/abs/1904.09051](http://arxiv.org/abs/1904.09051)
##### PDF
[http://arxiv.org/pdf/1904.09051](http://arxiv.org/pdf/1904.09051)
|
2024-11-18T18:20:11.944926+00:00 | 2020-11-17T16:47:04 | 327830c406bac847cf9c5e843340f89a9f3fb7a0 | 3 | {
"blob_id": "327830c406bac847cf9c5e843340f89a9f3fb7a0",
"branch_name": "refs/heads/master",
"committer_date": "2020-11-17T16:47:04",
"content_id": "0c3e2690f5e8abc803b3d40b7886cf7e6ffab91f",
"detected_licenses": [
"BSD-3-Clause"
],
"directory_id": "d594688faa447c8454fadd8fcf79dddaac7d4502",
"extension": "md",
"filename": "README.md",
"fork_events_count": 6,
"gha_created_at": "2016-10-04T15:26:04",
"gha_event_created_at": "2020-11-16T14:47:05",
"gha_language": "Python",
"gha_license_id": "BSD-3-Clause",
"github_id": 69978684,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 4221,
"license": "BSD-3-Clause",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:27",
"repo_name": "terraref/extractors-hyperspectral",
"revision_date": "2020-11-17T16:47:04",
"revision_id": "782e3c62088257bb9c8f0f4f1d8943fa02881dc0",
"snapshot_id": "889b50f9b3677dbd7012c5832a3607a3f3a97d11",
"src_encoding": "UTF-8",
"star_events_count": 6,
"url": "https://raw.githubusercontent.com/terraref/extractors-hyperspectral/782e3c62088257bb9c8f0f4f1d8943fa02881dc0/README.md",
"visit_date": "2021-07-05T06:19:21.297927"
} | 3.296875 | stackv2 | # Hyperspectral extractors
Date: 2017-11-13
## Authors
## Extractor Description
This repository contains extractors that process data originating from:
- Hyperspec INSPECTOR SWIR camera
- Hyperspec INSPECTOR VNIR camera
## Inputs and Outputs
This extractor processes ENVI BIL (band-interleaved-by-line) files into netCDF.
_Input_
- Evaluation is triggered whenever a file is added to a dataset
- Checks whether the file is a _raw file
_Output_
- The dataset containing the _raw file will get a corresponding .nc netCDF file
## Algorithm Description
### Theoretical basis
The hyperspectral calibration procedure is documented at https://github.com/terraref/computing-pipeline/issues/282. The implementation is undergoing improvement (e.g., better target calibrations, more reliable factory calibrations, improved interpolation methods) as experience is gained.
### Limitations
1. Only valid for 300-800nm (range of downwelling radiometer)
2. Raw hyperspectral exposures are calibrated for images in full sunlight. Other times and cloudy days are not well-tested yet.
3. Zenith angles used in the calibration data are: 42.3, 47.6, 53.0, 58.4, 64.0, 75.1, 80.7, 86.5
* The closer the raw data is to these angles the better the result
4. The calibration needs improvement to obtain accurate absolute reflectances. Environmental conditions (such as shade and specular reflection) and irregular calibration of known targets (such as tilted surfaces and field-based calibration) can bias retrieved reflectances by an unknown factor. However, the scale bias factors out of indices created as ratios of reflectances, e.g., (A-B) / (A+B).
## Application
### Files:
1. hyperspectral_workflow.sh
This is the main shell script:
- -c dfl_lvl Compression level [0..9] (empty means none) (default )
- -d dbg_lvl Debugging level (default 0)
- -h Create indices file. This has the same root name as out_fl but with the suffix "_ind.nc"
- -I drc_in Input directory (empty means none) (default )
- -i in_fl Input filename (required) (default )
- -j job_nbr Job simultaneity for parallelism (default 6)
- -m msk_fl location of Netcdf Soil Mask (Level 1 data) applied when creating indices file
- -n nco_opt NCO options (empty means none) (default )
- -N ntl_out Interleave-type of output (default bsq)
- -O drc_out Output directory (default /home/butowskh/terraref/extractors-hyperspectral/hyperspectral)
- -o out_fl Output-file (empty derives from Input filename) (default )
- -p par_typ Parallelism type (default bck)
- -t typ_out Type of netCDF output (default NC_USHORT)
- -T drc_tmp Temporary directory (default /gpfs_scratch/arpae/imaging_spectrometer)
- -u unq_sfx Unique suffix (prevents intermediate files from sharing names) (default .pid140080)
- -x xpt_flg Experimental (default No)
2. CalculationWorks.py
A supporting module for EnvironmentalLoggerAnalyser.py and JsonDealer.py.
This module is in charge of all the calculation works needed in the
EnvironmentalLoggerAnalyser.py (converting the data made by environmental logger)
and JsonDealer.py (group up the supporting files for data_raw).
* EnvironmentalLoggerAnalyzer.py
This module will read data generated by Environmental Sensor and convert to netCDF file
* JsonDealer.py
This module parses JSON formatted metadata and data and header provided by LemnaTec and outputs a formatted netCDF4 file
* DataProcess.py
This module will process the data file and export a netCDF with variables
from it and dimesions (band, x, y) from its hdr file
* hyperspectral_calibration.nco
NCO/ncap2 script to process and calibrate TERRAREF exposure data
## Failure Conditions
## Related GitHub issues and documentation
1. [Notes from meeting on calibration options](https://docs.google.com/document/d/e/2PACX-1vRKArTMn0aU90KoFKe-HCYMuFubcW_WLUZsFCWCT2rENhitzf00tLktYm6EG2DIB3X5rSRD1A1DOZhL/pub)
2. First (alpha) calibration proceedure https://github.com/terraref/computing-pipeline/issues/88
3. Second (radiometer based) calibration
* List of tasks https://github.com/terraref/computing-pipeline/issues/281
* Algorithm documentation https://github.com/terraref/computing-pipeline/issues/282
|
2024-11-18T18:20:11.989956+00:00 | 2020-10-01T15:14:52 | 87ad1c39cb5cf2fd995549c894af5cd7b001850d | 3 | {
"blob_id": "87ad1c39cb5cf2fd995549c894af5cd7b001850d",
"branch_name": "refs/heads/master",
"committer_date": "2020-10-01T15:14:52",
"content_id": "dbf306a25c9f44f99f3acec5f3dbcaba08714444",
"detected_licenses": [
"BSD-2-Clause"
],
"directory_id": "67eee1451b49759416d7f13de473b71bc0614d20",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": "2020-10-01T15:12:03",
"gha_event_created_at": "2020-10-01T15:12:03",
"gha_language": null,
"gha_license_id": "BSD-2-Clause",
"github_id": 300322971,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 3489,
"license": "BSD-2-Clause",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:28",
"repo_name": "dinithiravi/c",
"revision_date": "2020-10-01T15:14:52",
"revision_id": "953228b9dc7bdc08f23876c55453164e26050631",
"snapshot_id": "8477d60bc21368c01d496d55b48584f854e10193",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/dinithiravi/c/953228b9dc7bdc08f23876c55453164e26050631/README.md",
"visit_date": "2022-12-23T22:14:43.375735"
} | 3.390625 | stackv2 | # A small, fast C compiler suite.
NOTE - This project is not being actively developed. Please direct yourself to https://github.com/michaelforney/cc which
is a successor that is more complete. I direct my own fixes to that project instead now.
[](https://gitter.im/andrewchambers/c?utm_source=badge&utm_medium=badge&utm_campaign=pr-badge&utm_content=badge)
- Small.
- Fast.
- Consistent.
- High quality.
- Low complexity.
- No dependencies.
- No fussy configuration.
- Painless cross compiling.
- Just work.
You should be able to get a C compiler, assembler, linker and libc for any
supported target in less than 30 seconds.
## Building
Requires an external C compiler and gnu binutils (for now), and I have only tested
it on linux 64 bit so far.
The code does use anonymous union extensions, so your compiler will need to support them too.
```
$ make
```
## Testing
```
$ make test
$ make selfhost # self hosting
```
## Plan
### Stage 1.
### Stage 2.
Self hosting arm, something like raspberry pi/android.
### Stage 3.
Build small clean C code bases like 8cc, tcc, sbase.
### Stage 4.
Build musl libc.
### Beyond.
- Build more programs.
- Replace gnu as with our own assembler.
- Replace ld with our own static linker.
- Build OS kernels.
- SSA backend.
## Status
Pre stage 2. Self hosting with lots of missing common cases. Though technically these
bugs can be fixed with the compiler itself :). It uses it's own stubbed out headers
and cannot correctly process system headers yet (Help wanted).
## Contributing
Project on hold. See https://github.com/michaelforney/cc for a new compiler project that is more complete.
### Code layout
- Libraries are in src/*
- Commands are in src/cmd/*
If you are unsure about the purpose of a library, check the header which
should give a short description.
### Code style
Follow Plan9 style conventions. Headers are not allowed to include
other headers to eliminate circular dependencies and increase build speed.
src/u.h is the only exception to this rule.
- http://www.lysator.liu.se/c/pikestyle.html
- http://plan9.bell-labs.com/magic/man2html/6/style
- http://aiju.de/b/style
### Bug fixes and issues
Try and attach a single source file which exibits your issue. If possible
reduce the test case by hand until it is as small as possible.
Try and follow the general template changed where needed:
```
What are you trying to do:
...
What you expected to happen:
...
What actually hapened:
...
```
Try and add a small self contained file which reproduces the issue.
In general each bug fix or change should add a test file which triggers the bug.
### Memory management
The compiler does not explicitly free memory. Peak memory usage while self hosting
is approximately 2Mb, so it should not be an issue, even for planned targets/hosts like
the raspberry pi.
This actually simplifies the code and probably makes it faster because allocations can be pointer bumps.
## Useful Links
- C11 standard final draft http://www.open-std.org/jtc1/sc22/wg14/www/docs/n1570.pdf
- Dave Prosser's C Preprocessing Algorithm http://www.spinellis.gr/blog/20060626/
- The x86-64 ABI http://www.x86-64.org/documentation/abi.pdf
- http://aiju.de/rant/cross-compiling
- http://bellard.org/tcc/
- https://github.com/rui314/8cc
- http://harmful.cat-v.org/software/
- http://suckless.org/philosophy
|
2024-11-18T18:20:12.127399+00:00 | 2015-02-07T11:33:20 | cc2c566a975b5b0299080ad85db1a58c5f96a4cf | 2 | {
"blob_id": "cc2c566a975b5b0299080ad85db1a58c5f96a4cf",
"branch_name": "refs/heads/master",
"committer_date": "2015-02-07T11:33:20",
"content_id": "888377ce6cc2dc3501a5e589cc2f907171202245",
"detected_licenses": [
"MIT"
],
"directory_id": "65960a237f97fcbbcaf13c2d9e8665fa8ad2952d",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": 30453851,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 168,
"license": "MIT",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:29",
"repo_name": "githuntio/Programming-Languages",
"revision_date": "2015-02-07T11:33:20",
"revision_id": "c6621ed5b83383256a2dc2c43a1a0ba35fd70f88",
"snapshot_id": "30bcce1553c18ee49b68d775a1213ca2729b2f37",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/githuntio/Programming-Languages/c6621ed5b83383256a2dc2c43a1a0ba35fd70f88/README.md",
"visit_date": "2021-05-16T02:50:36.391706"
} | 2.03125 | stackv2 | ## Programming Languages
Those common programming languages in json format. [take a look](languages.json)
Remind me if I missed something.
## License
[MIT](LICENSE) |
2024-11-18T18:20:12.165658+00:00 | 2018-06-08T07:54:01 | 1a1121e1b7b00a13e23801baee8a95c8480c5e04 | 3 | {
"blob_id": "1a1121e1b7b00a13e23801baee8a95c8480c5e04",
"branch_name": "refs/heads/master",
"committer_date": "2018-06-08T07:54:01",
"content_id": "ea97f2459efc2a39998e818a8810df37cf70044b",
"detected_licenses": [
"MIT"
],
"directory_id": "bc661e599d4d91b7819e6278d3b84e707d7e29da",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": "2018-02-23T22:42:47",
"gha_event_created_at": "2018-06-08T07:54:02",
"gha_language": "Python",
"gha_license_id": "MIT",
"github_id": 122680300,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 3859,
"license": "MIT",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:30",
"repo_name": "schwerpunkt/zfsDiffReport.py",
"revision_date": "2018-06-08T07:54:01",
"revision_id": "0e171f220a12fa011e5607569e0a5886f9ef8903",
"snapshot_id": "6b3d7e45f03aa5594161d2ba66141dab3569ce5d",
"src_encoding": "UTF-8",
"star_events_count": 3,
"url": "https://raw.githubusercontent.com/schwerpunkt/zfsDiffReport.py/0e171f220a12fa011e5607569e0a5886f9ef8903/README.md",
"visit_date": "2021-04-27T01:42:46.696058"
} | 3.421875 | stackv2 | # zfsDiffReport.py
## You have...
- all your data in ZFS on your home server
- continuous snapshot routines of your volumes (e.g. with [zfs-auto-snapshot](https://github.com/zfsonlinux/zfs-auto-snapshot))
- [optional] complete trust in your data being safe from bitrot, ransomware... because you have a good mirror/send-recv backup routine for your snapshots
- ...however: *noticeable paranoia* that a file or files you deeply care about are unintentionally removed or modified by you without you noticing and you are longing for a simple python script that reports exactly that information
Well...
*zfsDiffReport.py* is a python script that diffs the latest two snapshots of a ZFS volume and prints report text files.
## Suggested usecase:
- set up [zfs-auto-snapshot](https://github.com/zfsonlinux/zfs-auto-snapshot) for weekly snapshots with e.g. the identifier 'zas_w-utc' (actually do it more frequently... something reasonable... four times an hour or so)
- set up a weekly cronjob and call zfsDiffReport like in the usage below (without -h)
- now, once every week, good time is a Monday, when you read your morning paper in your left hand, hold the spoon for your cereal in your right, and scroll down your phone with your nose, you might as well click on the *\*_zfsDiffReport.txt* file on your server and check the short list of changes you intentionally or accidentally made during the previous week
And that's how I think you will never lose any data.
## Usage:
```
./zfsDiffReport.py zfsTestPool/VolOne --debug -u philipp -f -r -s freq -s zas_f -h
usage: zfsDiffReport.py [-h] [-s [SNAPSHOTKEYS [SNAPSHOTKEYS ...]]]
[-o OUTDIR] [-f [FILENAME]]
[--outfilesuffix OUTFILESUFFIX] [-u USER] [-e EXCLUDE]
[-r] [--zfsbinary ZFSBINARY] [--debug] [-q]
volume [volume ...]
zfsDiffReport.py generates a report text file from the ZFS diff of a given
volume's two last snapshots containing a given identifier. The script is
intended to be used as companion to zfs-auto-snapshot. I use it to check my
weekly snapshots for unintended file deletions.
positional arguments:
volume observed ZFS volume(s) e.g.: 'ZPOOL/ZFSVOL'
optional arguments:
-h, --help show this help message and exit
-s [SNAPSHOTKEYS [SNAPSHOTKEYS ...]], --snapshotkeys [SNAPSHOTKEYS [SNAPSHOTKEYS ...]]
snapshot keywords e.g.: 'zas_w-utc-', no keyword: diff
latest snapshots, one keyword: diff latest snapshots
containing the keyword, two keywords: diff latest
snapshots each containing given keywords respectively
-o OUTDIR, --outdir OUTDIR
report file output directory
-f [FILENAME], --filename [FILENAME]
if not set each volume diff is written to a separate
file. if set all volume diffs are written to it, if
empty all reports are written to stdout, if not set
one report per volume is created
--outfilesuffix OUTFILESUFFIX
suffix for report text file; default:
'_zfsDiffReport.txt'
-u USER, --user USER user for output file e.g.: 'user'
-e EXCLUDE, --exclude EXCLUDE
multiple definitions possible; diff lines containing
an exclude keyword will be omitted e.g. '.git'
-r, --reduce ZFS lists a file that is deleted and (re)created
between snapshots with - and +; omit those lines when
the files' checksums match
--zfsbinary ZFSBINARY
path to ZFS binary; default: 'zfs'
--debug
-q, --quiet
And that's how you report a ZFS diff.
```
|
2024-11-18T18:20:12.456153+00:00 | 2023-04-27T19:58:08 | 463bd83e66260bd40aca623e2407cbe1b4564650 | 3 | {
"blob_id": "463bd83e66260bd40aca623e2407cbe1b4564650",
"branch_name": "refs/heads/master",
"committer_date": "2023-04-27T19:58:08",
"content_id": "85dd73624b01d281b45e837af21160367860b67b",
"detected_licenses": [
"MIT"
],
"directory_id": "967bcff6fa4bc38e75fb2e7dd36f0f52eaf1b902",
"extension": "md",
"filename": "README.md",
"fork_events_count": 8,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": 154877045,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 1509,
"license": "MIT",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:31",
"repo_name": "RonnyWeiss/Material-Dashboard-Cards-as-HTML-or-for-Apex",
"revision_date": "2023-04-27T19:58:08",
"revision_id": "6c2743fbaf7402ff74961109fefa01bf8550ba14",
"snapshot_id": "aa842f3744f5f66067b3ca42a2ad13ef51e0d0ad",
"src_encoding": "UTF-8",
"star_events_count": 23,
"url": "https://raw.githubusercontent.com/RonnyWeiss/Material-Dashboard-Cards-as-HTML-or-for-Apex/6c2743fbaf7402ff74961109fefa01bf8550ba14/README.md",
"visit_date": "2023-05-15T02:40:13.768723"
} | 2.8125 | stackv2 | # Material-Dashboard-Cards-as-HTML-or-for-Apex
  

This project is a Region Plug-in for Oracle APEX that let's you easily create some nice, responsive Material Design Cards. These Cards can be shown in a small view with just an icon or in a larger view with Chartist Charts. You can also use the cards without APEX. Just download all files and start index.html.
The Cards are data driven so you don't need to worry about the UI presentation. You only have to change the JSON in "data\data.js".
When you use these Cards in Apex, the complete help and some sample statements are inside of the Region Plug-in.
if you don't to know how to install this Plug-in in Apex, please take look at the API of Oracle Apex.
For working Demo just click on:
https://apex.oracle.com/pls/apex/f?p=103428
If you like my stuff, donate me a coffee
[](https://www.paypal.me/RonnyW1)
**Important clarification: My work in the development team of Oracle APEX is in no way related to my open source projects or the plug-ins on apex.world! All plug-ins are built in my spare time and are not supported by Oracle!** |
2024-11-18T18:20:12.534136+00:00 | 2018-10-22T13:20:09 | a0ee26ef3c8137b4f7a4c35cd8d047c787dc214d | 2 | {
"blob_id": "a0ee26ef3c8137b4f7a4c35cd8d047c787dc214d",
"branch_name": "refs/heads/master",
"committer_date": "2018-10-22T13:20:09",
"content_id": "16071c9559e2cdf42ffb1f2193c221ee0657da42",
"detected_licenses": [
"MIT"
],
"directory_id": "d45c7a15a5edf6a43a6a047c2ee85b60fed73d0e",
"extension": "md",
"filename": "README.md",
"fork_events_count": 2,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": 139801774,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 391,
"license": "MIT",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:32",
"repo_name": "TeamASMR/WebVoiceChat-with-ASRNNoise",
"revision_date": "2018-10-22T13:20:09",
"revision_id": "4cc56c7ad66ab8998e230693bab3df6e11ebd54e",
"snapshot_id": "246d81baac7536ad4bd5147e7cad21fbd88f9f50",
"src_encoding": "UTF-8",
"star_events_count": 10,
"url": "https://raw.githubusercontent.com/TeamASMR/WebVoiceChat-with-ASRNNoise/4cc56c7ad66ab8998e230693bab3df6e11ebd54e/README.md",
"visit_date": "2020-03-22T08:55:58.124193"
} | 2.1875 | stackv2 | # WebVoiceChat-with-ASRNNoise
* Web based voice talk application using Audio Super Resolution and RNNoise(called ASRNNoise).
* PNU-CSE-2018-graduate assignment
### References
* Fast-Audio-Super-Resolution - https://github.com/leekh7411/Fast-Audio-Super-Resolution.git
* RNNoise-Project - https://github.com/xiph/rnnoise
* RTCMulticonnection - https://github.com/muaz-khan/RTCMultiConnection |
2024-11-18T18:20:12.570792+00:00 | 2023-08-31T00:00:11 | 39603137cf466d6f41b6de4d1b7a3dc98e022f89 | 4 | {
"blob_id": "39603137cf466d6f41b6de4d1b7a3dc98e022f89",
"branch_name": "refs/heads/master",
"committer_date": "2023-08-31T00:00:11",
"content_id": "d6dfb8b9953006ef857e3a2536f9a4384d2f58ff",
"detected_licenses": [
"MIT",
"Apache-2.0"
],
"directory_id": "f60b94a7fa6db3ce08edd1c9324ac1cd66adadea",
"extension": "md",
"filename": "operations-scripting_content_rabbit_[,7).partial.md",
"fork_events_count": 393,
"gha_created_at": "2013-07-28T23:18:15",
"gha_event_created_at": "2023-09-14T20:00:11",
"gha_language": "C#",
"gha_license_id": "NOASSERTION",
"github_id": 11727301,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 2186,
"license": "MIT,Apache-2.0",
"license_type": "permissive",
"path": "/transports/rabbitmq/operations-scripting_content_rabbit_[,7).partial.md",
"provenance": "stack-edu-markdown-0000.json.gz:33",
"repo_name": "Particular/docs.particular.net",
"revision_date": "2023-08-31T00:00:11",
"revision_id": "800c70c539e1e46e5451f7e3224163cd0bb93566",
"snapshot_id": "0ea5bc851e24fc6fda58bf48a3bfdc6091fa5353",
"src_encoding": "UTF-8",
"star_events_count": 106,
"url": "https://raw.githubusercontent.com/Particular/docs.particular.net/800c70c539e1e46e5451f7e3224163cd0bb93566/transports/rabbitmq/operations-scripting_content_rabbit_[,7).partial.md",
"visit_date": "2023-08-31T14:09:46.373177"
} | 3.828125 | stackv2 | Example code and scripts to facilitate deployment and operational actions against RabbitMQ.
These samples use the [RabbitMQ.Client NuGet Package](https://www.nuget.org/packages/RabbitMQ.Client/).
Since the RabbitMQ.Client is not [CLS Compliant](https://msdn.microsoft.com/en-us/library/system.clscompliantattribute.aspx) it is not possible to run this code within PowerShell.
## Native Send
### The native send helper methods
A send involves the following actions:
* Create and serialize headers.
* Write a message body directly to RabbitMQ.
#### In C#
snippet: rabbit-nativesend
### Using the native send helper methods
snippet: rabbit-nativesend-usage
In this example, the value `MyNamespace.MyMessage` represents the .NET type of the message. See the [headers documentation](/nservicebus/messaging/headers.md) for more information on the `EnclosedMessageTypes` header.
## Return message to source queue
### The retry helper methods
This code shows an example of how to perform the following actions:
* Read a message from the error queue.
* Extract the failed queue from the headers.
* Forward that message to the failed queue name so it can be retried.
snippet: rabbit-return-to-source-queue
### Using the retry helper methods
snippet: rabbit-return-to-source-queue-usage
## Create queues
Queue creation can be done for a specific endpoint or queues shared between multiple endpoints.
### The create queue helper methods
snippet: rabbit-create-queues
### Creating queues for an endpoint
To create all queues for a given endpoint name.
snippet: rabbit-create-queues-for-endpoint
### Using the create endpoint queues
snippet: rabbit-create-queues-endpoint-usage
### To create shared queues
snippet: rabbit-create-queues-shared-usage
## Create HA policy
To configure HA policy, refer to the [RabbitMQ HA documentation](https://www.rabbitmq.com/ha.html).
## Delete queues
### The delete helper queue methods
snippet: rabbit-delete-queues
### To delete all queues for a given endpoint
snippet: rabbit-delete-queues-for-endpoint
snippet: rabbit-delete-queues-endpoint-usage
### To delete shared queues
snippet: rabbit-delete-queues-shared-usage
|
2024-11-18T18:20:15.344053+00:00 | 2018-03-28T17:09:44 | 408c03200b734308ffd9d3da1e0fa079c23f968d | 2 | {
"blob_id": "408c03200b734308ffd9d3da1e0fa079c23f968d",
"branch_name": "refs/heads/master",
"committer_date": "2018-03-28T17:09:44",
"content_id": "f6e46e4c0489d915c981bdfba626ff3f8b14fa80",
"detected_licenses": [
"MIT"
],
"directory_id": "f148154b05883a166974f85920ad188dcdf973ac",
"extension": "md",
"filename": "README.md",
"fork_events_count": 1,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": 126763981,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 271,
"license": "MIT",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:34",
"repo_name": "ericsk/azure-serverless-hol",
"revision_date": "2018-03-28T17:09:44",
"revision_id": "681294eb1434fc4ece6e8e51372e4842eda5fb6a",
"snapshot_id": "670aa06cbeaa96906274c9c0c54f2ede5452b875",
"src_encoding": "UTF-8",
"star_events_count": 2,
"url": "https://raw.githubusercontent.com/ericsk/azure-serverless-hol/681294eb1434fc4ece6e8e51372e4842eda5fb6a/README.md",
"visit_date": "2021-04-15T05:50:52.019100"
} | 2.171875 | stackv2 | # Azure Serverless Hands-on Labs.
* [Face Identification](Labs/FaceIdentificationLab.md): Use Azure Function to handle the face training and identification by leveraging [Azure Cognitive Services: Face API](https://azure.microsoft.com/services/cognitive-services/face/). |
2024-11-18T18:20:15.380374+00:00 | 2013-05-28T13:29:13 | c08faefa6682fee24124fd215e6a3c1e837d92da | 3 | {
"blob_id": "c08faefa6682fee24124fd215e6a3c1e837d92da",
"branch_name": "refs/heads/merged",
"committer_date": "2013-05-28T13:29:13",
"content_id": "57f48db4eea8a3803300da182e67963ea244109f",
"detected_licenses": [
"BSD-2-Clause"
],
"directory_id": "b7e536c28df0108774f27640be1f62babd68aae9",
"extension": "md",
"filename": "ex.md",
"fork_events_count": 0,
"gha_created_at": "2012-07-08T18:58:36",
"gha_event_created_at": "2013-05-28T13:29:14",
"gha_language": "Objective-C",
"gha_license_id": null,
"github_id": 4948909,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 430,
"license": "BSD-2-Clause",
"license_type": "permissive",
"path": "/help/ex.md",
"provenance": "stack-edu-markdown-0000.json.gz:35",
"repo_name": "Shadowfiend/vico",
"revision_date": "2013-05-28T13:29:13",
"revision_id": "5f8c1a2461e57e4c00e71529339d5cfc2b2d1e62",
"snapshot_id": "7e86220d56c23a9f8387b0437d443c20a130877e",
"src_encoding": "UTF-8",
"star_events_count": 2,
"url": "https://raw.githubusercontent.com/Shadowfiend/vico/5f8c1a2461e57e4c00e71529339d5cfc2b2d1e62/help/ex.md",
"visit_date": "2021-01-18T16:19:26.665939"
} | 3.0625 | stackv2 | # The ex command line
The ex command line is normally hidden under the status bar at the
bottom of the window. The command line is used to enter ex commands
(<kbd>:</kbd>), filter commands (<kbd>!</kbd>) and search patterns
(<kbd>/</kbd> and <kbd>?</kbd>).
Lines entered on the command line are remembered in a history. Use up and
down arrows to recall history.
* [Ex ranges](ex_ranges.html)
* [Ex commands](ex_cmds.html)
|
2024-11-18T18:20:15.507953+00:00 | 2021-01-19T14:04:49 | ff5bed0d16c2b609fccd12757887768f3561ac6b | 3 | {
"blob_id": "ff5bed0d16c2b609fccd12757887768f3561ac6b",
"branch_name": "refs/heads/main",
"committer_date": "2021-01-19T14:04:49",
"content_id": "cac75f66a9882a41b335816d7abb90a06f96c4de",
"detected_licenses": [
"MIT"
],
"directory_id": "95093a85dd390cdd0b1eaf5b0b65deea192c6802",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": 325562735,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 1422,
"license": "MIT",
"license_type": "permissive",
"path": "/vga_controller_160x120x1/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:36",
"repo_name": "MarcelMG/my_VHDL_playground",
"revision_date": "2021-01-19T14:04:49",
"revision_id": "99a8de6d33dbf7f813b8a4e350b5275fbf4b295d",
"snapshot_id": "0d2df6395bc171eadfa062a4e571d8c89e8c71d0",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/MarcelMG/my_VHDL_playground/99a8de6d33dbf7f813b8a4e350b5275fbf4b295d/vga_controller_160x120x1/README.md",
"visit_date": "2023-02-12T21:46:47.486796"
} | 3.3125 | stackv2 | # VGA controller written in VHDL
features:
* 160x120 pixel image resolution with 4x upscaling to 640x480 VGA signal
* 1bit black and white (monochrome) color depth
* BRAM frame buffer stores image to be displayed, can be modified at any time
how is the 160x120 image upscaled to 640x480 resolution?
-> as the VGA controller 'scans' the horizontal and vertical lines of the video signal, it keeps track of the current pixel position with a horizontal and a vertical pixel counter. A 2bit-counter is used to divide this count by 4, so we get two 'downscaled' pixel positions. To finally compute the linear adress of the pixel value stored in the BRAM frame buffer, we compute adress = 160 * vertical_pos + horizontal_pos. For the multiplication, a 9bit hardware multiplier of the MAX10 FPGA is used.
example top.vhdl:
a sample image is written to the memory during programming using a memory initialization file (.mif) this image is displayed by the monitor
the .mif-file was created with a simple Python script, you can find it [here](https://github.com/MarcelMG/Miscellaneous/blob/master/image_to_altera_mif/image_to_altera_mif.ipynb).
to simulate the design with ModelSim, go to the directory and execute 'do testbench.tcl'
the whole demo project for the MAX1000 development board is in the .qar-file

|
2024-11-18T18:20:18.570806+00:00 | 2020-05-19T20:37:38 | fe8d4c9559e9df299aed14bddab030b5dd28c553 | 4 | {
"blob_id": "fe8d4c9559e9df299aed14bddab030b5dd28c553",
"branch_name": "refs/heads/master",
"committer_date": "2020-05-19T20:37:38",
"content_id": "28022f6753ed353a9bb537bea14ca8771ac714ca",
"detected_licenses": [
"MIT"
],
"directory_id": "3f21e4fbdd846f71c968056dbcbbb44e234c900d",
"extension": "md",
"filename": "Poisson_main.md",
"fork_events_count": 1,
"gha_created_at": "2019-08-07T23:21:57",
"gha_event_created_at": "2020-09-04T17:09:38",
"gha_language": "HTML",
"gha_license_id": "NOASSERTION",
"github_id": 201141761,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 7399,
"license": "MIT",
"license_type": "permissive",
"path": "/Markdown/Poisson_main.md",
"provenance": "stack-edu-markdown-0000.json.gz:37",
"repo_name": "mehak-sachdeva/MGWR_book",
"revision_date": "2020-05-19T20:37:38",
"revision_id": "d702a04c7e5f9101e0f04b6d84295bce86579059",
"snapshot_id": "1ed31a1cbab3493a64aaf7cde36ede35fff743b9",
"src_encoding": "UTF-8",
"star_events_count": 1,
"url": "https://raw.githubusercontent.com/mehak-sachdeva/MGWR_book/d702a04c7e5f9101e0f04b6d84295bce86579059/Markdown/Poisson_main.md",
"visit_date": "2021-07-12T21:50:44.321259"
} | 3.640625 | stackv2 |
# <center>Multiscale Geographically Weighted Regression - Poisson dependent variable</center>
The model has been explored and tested for multiple parameters on real and simulated datasets. The research includes the following outline with separate notebooks for each part.
**Notebook Outline:**
**Introduction (current)**
- [Introduction](#Introduction)
- [Introduction to the problem](#Introduction-to-the-project)
- [Statistical Equations](#Statistical-Equations)
- [Approaches Explored](#Approaches-Explored)
- [Other notebooks](#Other-notebooks)
- [References](#References)
[Back to the main page](https://mehak-sachdeva.github.io/MGWR_book/)
---
# Introduction
***
## Introduction to the problem
A recent addition to the local statistical models in PySAL is the implementation of Multiscale Geographically Weighted Regression (MGWR) model, a multiscale extension to the widely used approach for modeling process spatial heterogeneity - Geographically Weighted Regression (GWR). GWR is a local spatial multivariate statistical modeling technique embedded within the regression framework that is calibrated and estimates covariate parameters at each location using borrowed data from neighboring observations. The extent of neighboring observations used for calibration is interpreted as the indicator of scale for the spatial processes and is assumed to be constant across covariates in GWR. MGWR, using a back-fitting algorithm relaxes the assumption that all processes being modeled operate at the same spatial scale and estimates a unique indicator of scale for each process.
The GWR model in PySAL can currently estimate Gaussian, Poisson and Logistic models though the MGWR model is currently limited to only Gaussian models. This project aims to expand the MGWR model to nonlinear local spatial regression modeling techniques where the response outcomes may be discrete (following a Poisson distribution). This will enable a richer and holistic local statistical modeling framework to model multi-scale process heterogeneity for the open source community.
## Statistical Equations
A conventional Poisson regression model is written as:
\begin{align}
O_i {\sim} Poisson[E_i exp ({\sum} {\beta} & _k x _{k,i})] \\
\end{align}
where $x_{k,1}$ is the kth explanatory variable in place i and the ${\beta}_ks$ are the parameters and Poisson indicates a Poisson distribution with mean $\lambda$.
Nakaya et.al. (2005) introduced the concept of allowing parameter values to vary with geographical location ($u_i$), which is a vector of two dimensional co-ordinates describing the location *i*. The Poisson model for geographically varying parameters can be written as:
\begin{align}
O_i {\sim} Poisson[E_i exp ({\sum} {\beta} & _k (u_i) x _{k,i})] \\
\end{align}
The Geographically Weighted Poisson Regression model (GWPR) is estimated using a modified local Fisher scoring procedure, a form of iteratively reweighted least squares (IRLS). In this procedure, the following matrix computation of weighted least squares should be repeated to update parameter estimates until they converge (Nakaya et.al., 2005):
\begin{align}
\beta^{(l+1)} (u_i) = (X^{t} W (u_i) A(u_i)^{(l)} X)^{-1} X^{t} W (u_i) A (u_i) ^{(l)} z (u_i){(l)} \\
\end{align}
# Approaches Explored
**Expected theoretical model calibration:**<br><br>
The calibration methodology for modeling response variables with a Poisson distribution in MGWR, through references from Geographically Weighted Poisson Regression (GWPR) (Nakaya, et al., 2005) and literature on Generalized Additive Models (Hastie & Tibshirani, 1986), is expected to be as follows:<br>
1. Initialize using GWPR estimates $𝛽_𝑘(𝑢_𝑖)^0: f^1_0,f^2_0, … , f^𝐾_0, $
where $f_𝑘^0 = (𝑥_{1 𝑘}𝛽_𝑘(𝑢_1)^0 … 𝑥_{𝑁𝑘}𝛽_𝑘(𝑢_𝑁)^0)$<br><br>
2. Update for each location ($𝑢_𝑖$) an adjusted dependent variable: <br>
$z(𝑢_𝑖)^{(𝑙)} = (z_1(𝑢_𝑖)^{(𝑙)}, z_2(𝑢_𝑖)^{(𝑙)}, … ,z_𝑁(𝑢_𝑖)^{(𝑙)})^𝑡$<br>
$z_𝑗(𝑢_𝑖)^{(𝑙)} = Σ𝛽_𝑘(𝑢_𝑖)^{(𝑙)} 𝑥_{𝑗,𝑘} + O_𝑗 − Ô_𝑗(𝛽(𝑢_𝑖)^{(𝑙)}) / Ô_𝑗(𝛽(𝑢_𝑖)^{(𝑙)})$<br><br>
3. Construct weights as follows:
$A(𝑢_𝑖)^{(𝑙)}$ which is a diagonal matrix with values ($Ô_1(𝛽(
𝑢_𝑖)^{(𝑙)}), Ô_2(𝛽(
𝑢_𝑖)^{(𝑙)}), ... ,Ô_𝑁(𝛽(𝑢_𝑖)^{(𝑙)}))$<br><br>
4. Fit an MGWR model to $z(𝑢_𝑖)^{(𝑙)}$ to update $𝛽_𝑘(𝑢_𝑖)$ and $z(𝑢_𝑖)^{(𝑙)}$, using the new weight matrix:<br>
$𝑊_𝑘∗(𝑢_𝑖)^{(𝑙)}=𝑊_𝑘(𝑢_𝑖)^{(𝑙)} A(𝑢_𝑖)^{(𝑙)}$<br><br>
5. Repeat steps (2) through (4) until convergence.
# Other notebooks
Below are links to the tests and exploration for the finalized MGWR model for Poisson dependent variables:
## Tests with real and simulated data
***
**[Initial module changes and univariate model check ](http://mehak-sachdeva.github.io/MGWR_book/Html/Poisson_MGWR_univariate_check)**
- Setup with libraries
- Fundamental equations for Poisson MGWR
- Example Dataset
- Helper functions
- Univariate example
- Parameter check
- Bandwidths check
**[Simulated Data example](http://mehak-sachdeva.github.io/MGWR_book/Html/Simulated_data_example_Poisson-MGWR)**
- Setup with libraries
- Create Simulated Dataset
- Forming independent variables
- Creating y variable with Poisson distribution
- Univariate example
- Bandwidth: Random initialization check
- Parameters check
- Multivariate example
- Bandwidths: Random initialization check
- Parameters check
- Global model parameter check
**[Real Data example](http://mehak-sachdeva.github.io/MGWR_book/Html/Real_data_example_Poisson-MGWR)**
- Setup with libraries
- Tokyo Mortality Dataset
- Univariate example
- Bandwidth: Random initialization check
- Parameter check
- Multivariate example
Bandwidths: Random initialization check
- MGWR bandwidths
- AIC, AICc, BIC check
## Monte Carlo Test
***
**[Monte Carlo Simulation Visualization](http://mehak-sachdeva.github.io/MGWR_book/Html/Poisson_MGWR_MonteCarlo_Results)**
- Setup with libraries
- List bandwidths from pickles
- Parameter functions
- GWR bandwidth
- MGWR bandwidths
- AIC, AICc, BIC check
- AIC, AICc, BIC Boxplots for comparison
- Parameter comparison from MGWR and GWR
# References
1. Fotheringham, A. S., Yang, W., & Kang, W. (2017). Multiscale Geographically Weighted Regression (MGWR). Annals of the American Association of Geographers, 107(6), 1247–1265. https://doi.org/10.1080/24694452.2017.1352480
2. Nakaya, T., Fotheringham, A. S., Brunsdon, C., & Charlton, M. (2005). Geographically weighted Poisson regression for disease association mapping. Statistics in Medicine, 24(17), 2695–2717. https://doi.org/10.1002/sim.2129
3. Yu, H., Fotheringham, A. S., Li, Z., Oshan, T., Kang, W., & Wolf, L. J. (2019). Inference in Multiscale Geographically Weighted Regression. Geographical Analysis, gean.12189. https://doi.org/10.1111/gean.12189
4. Hastie, T., & Tibshirani, R. (1986). Generalized Additive Models. Statistical Science, 1(3), 297–310. https://doi.org/10.1214/ss/1177013604
5. Wood, S. N. (2006). Generalized additive models : an introduction with R. Chapman & Hall/CRC.
[Back to the main page](https://mehak-sachdeva.github.io/MGWR_book/)
|
2024-11-18T18:20:18.607162+00:00 | 2022-06-27T12:57:49 | 68ce2c2ceb3ea747a7ebd598165b5d5d885c96e4 | 4 | {
"blob_id": "68ce2c2ceb3ea747a7ebd598165b5d5d885c96e4",
"branch_name": "refs/heads/master",
"committer_date": "2022-06-27T12:57:49",
"content_id": "0129ff39c4f9237c4b117e7be10637867769c020",
"detected_licenses": [
"Apache-2.0"
],
"directory_id": "88f6025447d3b1c5e44eb3e6b15b8d3372a8961a",
"extension": "md",
"filename": "README.md",
"fork_events_count": 6,
"gha_created_at": "2019-05-17T00:21:27",
"gha_event_created_at": "2022-06-27T10:25:34",
"gha_language": "Python",
"gha_license_id": "Apache-2.0",
"github_id": 187118831,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 5079,
"license": "Apache-2.0",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:38",
"repo_name": "opennetworkinglab/p4mn-docker",
"revision_date": "2022-06-27T12:57:49",
"revision_id": "fba9543e906129258eefa9826b9197a42b54dd38",
"snapshot_id": "29449f6dcf9d834b4091993e7945112f631321cc",
"src_encoding": "UTF-8",
"star_events_count": 16,
"url": "https://raw.githubusercontent.com/opennetworkinglab/p4mn-docker/fba9543e906129258eefa9826b9197a42b54dd38/README.md",
"visit_date": "2023-07-31T15:47:07.744693"
} | 3.90625 | stackv2 | # P4Runtime-enabled Mininet Docker Image
Docker image that can execute a Mininet-emulated network of BMv2 virtual
switches, controlled by an external SDN controller via P4Runtime.
This image was created to facilitate testing of P4Runtime support in the ONOS
controller, but you can use it without ONOS.
To obtain the image:
docker pull opennetworking/p4mn:<tag>
## Tags
The image comes in two versions (tags):
* `opennetworking/p4mn:latest` Updated daily and built from the master branch of
the [BMv2][BMv2] and [PI][PI] (P4Runtime server implementation) repositories;
* `opennetworking/p4mn:stable` Built using selected BMv2 and PI versions that
are known to work well with ONOS master.
Moreover, each tag is available in a "no-logging" version with disabled logging
macros in BMv2 to improve packet forwarding performance:
* `opennetworking/p4mn:latest-no-logging`
* `opennetworking/p4mn:stable-no-logging`
## Status [](https://github.com/opennetworkinglab/p4mn-docker/actions/workflows/main.yml)
Images are built daily using [Github Actions][GH Actions] and pushed to
[Docker Hub][Docker Hub].
## Steps to run p4mn
To run the container:
docker run --privileged --rm -it opennetworking/p4mn [MININET ARGS]
After running this command, you should see the mininet CLI (`mininet>`).
It is important to run this container in privileged mode (`--privileged`) so
mininet can modify the network interfaces and properties to emulate the desired
topology.
The image defines as entry point the mininet executable configured to use BMv2
`simple_switch_grpc` as the default switch. Options to the docker run command
(`[MININET ARGS]`) are passed as parameters to the mininet process. For more
information on the supported mininet options, please check the official mininet
documentation.
For example, to run a linear topology with 3 switches:
docker run --privileged --rm -it opennetworking/p4mn --topo linear,3
### P4Runtime server ports
Each switch starts a P4Runtime server which is bound to a different port,
starting from 50001 and increasing. To connect an external P4Runtime client
(e.g. an SDN controller) to the switches, you have to publish the corresponding
ports.
For example, when running a topology with 3 switches:
docker run --privileged --rm -it -p 50001-50003:50001-50003 opennetworking/p4mn --topo linear,3
### BMv2 logs and other temporary files
To allow easier access to BMv2 logs and other files, we suggest sharing the
`/tmp` directory inside the container on the host system using the docker run
`-v` option, for example:
docker run ... -v /tmp/p4mn:/tmp ... opennetworking/p4mn ...
By using this option, during the container execution, a number of files related
to the execution of the BMv2 switches will be available under `/tmp/p4mn` in the
host system. The name of these files depends on the switch name used in Mininet,
e.g. s1, s2, etc.
Example of these files are:
* `bmv2-s1-grpc-port`: contains the port used for the P4Runtime server executed
by the switch instance named `s1`;
* `bmv2-s1-log`: contains the BMv2 log;
* `bmv2-s1-netcfg.json`: ONOS netcfg JSON file that can be pushed to ONOS
to discover this switch instance. This file assumes that ONOS is running on
the same host as the container. If this is not the case, you will need to
modify the `managementAddress` property in the JSON file, replacing
`localhost` with the IP address of the host system of the `p4mn` container.
### Bash alias
A convenient way to quickly start the p4mn container is to create an alias in
your bash profile file (`.bashrc`, `.bash_aliases`, or `.bash_profile`) . |
2024-11-18T18:20:18.634054+00:00 | 2019-02-19T14:32:28 | 42a9d7fdcf66976b939f050fc918020561dc8ab5 | 3 | {
"blob_id": "42a9d7fdcf66976b939f050fc918020561dc8ab5",
"branch_name": "refs/heads/master",
"committer_date": "2019-02-19T14:32:28",
"content_id": "324becc3d38ee40f5bfae265692c32ff603f9b75",
"detected_licenses": [
"MIT"
],
"directory_id": "7180f367bd276178417657b9f739bf24fece56b0",
"extension": "md",
"filename": "README.md",
"fork_events_count": 1,
"gha_created_at": "2018-10-15T17:45:48",
"gha_event_created_at": "2018-10-15T17:45:48",
"gha_language": null,
"gha_license_id": null,
"github_id": 153156745,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 559,
"license": "MIT",
"license_type": "permissive",
"path": "/tests/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:39",
"repo_name": "Kukunin/aurora.js",
"revision_date": "2019-02-19T14:32:28",
"revision_id": "3b398c1dee5a7a9819562edc3691b1208b3e2f68",
"snapshot_id": "737aa668a6d936568f54bb52330d7922783e5f05",
"src_encoding": "UTF-8",
"star_events_count": 2,
"url": "https://raw.githubusercontent.com/Kukunin/aurora.js/3b398c1dee5a7a9819562edc3691b1208b3e2f68/tests/README.md",
"visit_date": "2020-04-01T11:18:53.488275"
} | 2.734375 | stackv2 | Tests
=====
The tests for Aurora are written using the [Mocha](http://mochajs.org/) testing framework. They
run in both Node.js and the browser.
## Setup
First, you'll need the test data, so init your git submodules to download them, and update them
if you've already downloaded them before.
git submodule init
git submodule update
## Running
1. Follow the setup steps above.
2. Run `make test` to test in both Node and PhantomJS.
3. Alternatively, you can run just `make test-browser` or `make test-node`
to run only in that environment.
|
2024-11-18T18:16:30.750123+00:00 | 2023-06-07T06:10:33 | 823e1112e7d01c45144a0cb0413be290ae9da56c | 3 | {
"blob_id": "823e1112e7d01c45144a0cb0413be290ae9da56c",
"branch_name": "refs/heads/master",
"committer_date": "2023-06-07T06:10:33",
"content_id": "629a312cb7853b1bcf4740fd40301e1c4cdf8fbe",
"detected_licenses": [
"BSD-2-Clause",
"Apache-2.0"
],
"directory_id": "d512f2cc5d795e74d629f63cbce2ec6c44894114",
"extension": "mdown",
"filename": "tasks.mdown",
"fork_events_count": 87,
"gha_created_at": "2014-11-17T05:22:59",
"gha_event_created_at": "2023-06-07T05:45:18",
"gha_language": "Makefile",
"gha_license_id": "Apache-2.0",
"github_id": 26741362,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 1558,
"license": "BSD-2-Clause,Apache-2.0",
"license_type": "permissive",
"path": "/rfc/src/tasks.mdown",
"provenance": "stack-edu-markdown-0000.json.gz:40",
"repo_name": "jmapio/jmap",
"revision_date": "2023-06-07T06:10:33",
"revision_id": "bdb932e26c0d729cbb351acde090b4903e28252a",
"snapshot_id": "411065ee28beb4b2d0f04b898918a4acc07d0cfc",
"src_encoding": "UTF-8",
"star_events_count": 1105,
"url": "https://raw.githubusercontent.com/jmapio/jmap/bdb932e26c0d729cbb351acde090b4903e28252a/rfc/src/tasks.mdown",
"visit_date": "2023-07-31T18:23:28.393438"
} | 2.53125 | stackv2 | %%%
title = "JMAP for Tasks"
abbrev = "JMAP Tasks"
category = "std"
docName = "draft-ietf-jmap-tasks-06"
ipr = "trust200902"
area = "Applications"
workgroup = "JMAP"
keyword = ["JMAP", "JSON", "tasks"]
date = 2023-03-10T00:00:00Z
[seriesInfo]
name="Internet-Draft"
value="draft-ietf-jmap-tasks-06"
stream="IETF"
status="standard"
[[author]]
initials="J.M."
surname="Baum"
fullname="Joris Baum"
role="editor"
organization = "audriga"
[author.address]
email =<EMAIL_ADDRESS> uri = "https://www.audriga.com"
[author.address.postal]
street = "Alter Schlachthof 57"
city = "Karlsruhe "
code = "76137"
country = "Germany"
[[author]]
initials="H.J."
surname="Happel"
fullname="Hans-Joerg"
role="editor"
organization = "audriga"
[author.address]
email =<EMAIL_ADDRESS> uri = "https://www.audriga.com"
[author.address.postal]
street = "Alter Schlachthof 57"
city = "Karlsruhe "
code = "76137"
country = "Germany"
%%%
.# Abstract
This document specifies a data model for synchronizing task data with a server using JMAP.
{mainmatter}
{{spec/tasks/intro.mdown}}
{{spec/tasks/principal.mdown}}
{{spec/tasks/list.mdown}}
{{spec/tasks/task.mdown}}
{{spec/tasks/tasknotifications.mdown}}
{{spec/tasks/securityconsiderations.mdown}}
{{spec/tasks/ianaconsiderations.mdown}}
{backmatter}
|
2024-11-18T18:16:30.877679+00:00 | 2021-09-20T07:47:09 | f5bb77bf5545fd5fd305f5480fabcc92b6864526 | 3 | {
"blob_id": "f5bb77bf5545fd5fd305f5480fabcc92b6864526",
"branch_name": "refs/heads/main",
"committer_date": "2021-09-20T07:47:09",
"content_id": "365eb4f19b70117e5c8fe72999326a976f7bfd39",
"detected_licenses": [
"MIT"
],
"directory_id": "cb15f99790deb0c2afff3347cecb1be68a267e2a",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": 408337955,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 503,
"license": "MIT",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:41",
"repo_name": "frostming/vscode-mingshe",
"revision_date": "2021-09-20T07:47:09",
"revision_id": "5af18ff9597393cae62d8bf0105af0b3ff5e02f4",
"snapshot_id": "03ba9701359d692b0fd36cf29a99e26959c19a0c",
"src_encoding": "UTF-8",
"star_events_count": 2,
"url": "https://raw.githubusercontent.com/frostming/vscode-mingshe/5af18ff9597393cae62d8bf0105af0b3ff5e02f4/README.md",
"visit_date": "2023-07-30T13:11:12.737642"
} | 2.84375 | stackv2 | # MíngShé Extension for VSCode
The language support of [MíngShé](https://github.com/abersheeran/mingshe) for VSCode.
## Features
- Syntax Highlight
## Requirements
- [VSCode-Python](https://marketplace.visualstudio.com/items?itemName=ms-python.python)
## Known Issues
## Credits
- The Python syntax highlight is based on [TextMate's Python Support](https://github.com/textmate/python.tmbundle)
## Release Notes
### 0.0.1
Initial release of MíngShé, add syntax highlighting.
**Enjoy!**
|
2024-11-18T18:03:44.752113+00:00 | 2021-08-22T13:59:06 | d1a1ab9476b9a35f32e92488d9fbc7550bde4136 | 2 | {
"blob_id": "d1a1ab9476b9a35f32e92488d9fbc7550bde4136",
"branch_name": "refs/heads/master",
"committer_date": "2021-08-22T13:59:06",
"content_id": "4ef198512c772fffed43ce2664c7458c56c17bb3",
"detected_licenses": [
"MIT"
],
"directory_id": "0efc4688b34677650ff245461dab438d34cbf7ca",
"extension": "md",
"filename": "README.md",
"fork_events_count": 2,
"gha_created_at": "2016-07-31T12:31:27",
"gha_event_created_at": "2021-08-22T13:59:30",
"gha_language": "Java",
"gha_license_id": "MIT",
"github_id": 64592868,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 836,
"license": "MIT",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:42",
"repo_name": "ravthiru/dictionary-store",
"revision_date": "2021-08-22T13:59:06",
"revision_id": "b94509af5b3d1a7e85f2dd3e1f7b5ec9fe0449bc",
"snapshot_id": "beab7dd97aa54b5fd58a4988d0f36d6b10008a60",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/ravthiru/dictionary-store/b94509af5b3d1a7e85f2dd3e1f7b5ec9fe0449bc/README.md",
"visit_date": "2021-08-28T14:26:51.620716"
} | 2.4375 | stackv2 | # Dictionary-store
While solving a word puzzle, i generate words with all combinations of the alphabets(a-z)
and filter them if it is a dictionary word. So i want a data-structure
to store english dictionary words and then perform a look-up to check if it is a dictionary word.
There are diffirent options to store word-list as Strings for easy look-up like Hashmap, Trie.
While searching I found [other option](http://stackoverflow.com/questions/2276641/way-to-store-a-large-dictionary-with-low-memory-footprint-fast-lookups-on-and) converting the word-list to Long's then searching using binary search. It consumes less memory when copared to other options.
This should work better than search trees as the dictionary is constant and binary search should be enough. Search trees sould be used when we need to modify the data.
|
2024-11-18T18:03:44.844434+00:00 | 2023-08-21T17:38:59 | 496471f78a30d9c5f37f0670aeaad0dc4a6f778c | 3 | {
"blob_id": "496471f78a30d9c5f37f0670aeaad0dc4a6f778c",
"branch_name": "refs/heads/master",
"committer_date": "2023-08-21T17:38:59",
"content_id": "ae1beaded4b83d9f55a6b7ccaa9be3cc71e2e2c6",
"detected_licenses": [
"MIT"
],
"directory_id": "58fe88a3cf0e448b0b31b14a47c006b55d236270",
"extension": "md",
"filename": "acknowledgements.md",
"fork_events_count": 3,
"gha_created_at": "2019-10-25T11:17:04",
"gha_event_created_at": "2022-07-28T20:20:18",
"gha_language": "Jinja",
"gha_license_id": "MIT",
"github_id": 217517401,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 50422,
"license": "MIT",
"license_type": "permissive",
"path": "/docs/acknowledgements.md",
"provenance": "stack-edu-markdown-0000.json.gz:43",
"repo_name": "bactopia/bactopia.github.io",
"revision_date": "2023-08-21T17:38:59",
"revision_id": "e43e2dea147db35c57477a40e69663bbf47cd7e0",
"snapshot_id": "e646a1e7c95735e27206282647489ad4101d0e7a",
"src_encoding": "UTF-8",
"star_events_count": 2,
"url": "https://raw.githubusercontent.com/bactopia/bactopia.github.io/e43e2dea147db35c57477a40e69663bbf47cd7e0/docs/acknowledgements.md",
"visit_date": "2023-09-01T07:03:36.011579"
} | 3.328125 | stackv2 | ---
title: Acknowledgements
description: >-
A full list of the 144 datasets and software packages used by
Bactopia, along with links and citations.
---
# Acknowledgements
Bactopia is truly a case of *"standing upon the shoulders of giants"*. Bactopia
currently includes more than 144 datasets and software packages. Nearly
every component uses, from the workflow, datasets, software packages, even the
framework for this site, was created by others and made freely available to the public.
I would like to personally extend my many thanks and gratitude to the authors
of these software packages and public datasets.
Really, thank you very much!
!!! info "Please Cite Datasets and Tools"
If you have used Bactopia in your work, please be sure to cite any datasets
or software you may have used.
## Funding
Support for this project came (in part) from an Emory Public Health Bioinformatics Fellowship
funded by the CDC Emerging Infections Program (U50CK000485) PPHF/ACA: Enhancing Epidemiology
and Laboratory Capacity.
## Influences
### nf-core
[nf-core](https://nf-co.re/) is group of great individuals volunteering
their time to create a set of curated Nextflow analysis pipelines. The [nf-core Team](https://nf-co.re/about)
is putting together some amazing practices that I think really strengthen the
Nextflow community as a whole!
Since the beginning I have always had the idea of Bactopia one day being
apart of nf-core. Unfortunately, I think Bactopia is a much to big at this point to fit the nf-core mold, but that's OK!
There are still many nf-core practices I've been able to adapt into Bactopia.
1. Arguement parsing based on nf-core library
2. All Bactopia Tools are adapted from nf-core/modules
3. Testing implemented to follow nf-core/modules
By implementing these practices, Bactopia I believe is much better pipeline to use. For this I'm very
grateful to the nf-core community! Thank you!
Ewels P, Peltzer A, Fillinger S, Patel H, Alneberg J, Wilm A, Garcia MU, Di Tommaso P, Nahnsen S [The nf-core framework for community-curated bioinformatics pipelines.](https://dx.doi.org/10.1038/s41587-020-0439-x) _Nat Biotechnol._ (2020)
## Public Datasets
Below is a list of 19 public datasets that could have potentially
been included during the *[Build Datasets](datasets.md)* step.
### Ariba Reference Datasets
These datasets are available using Ariba's `getref` function. You can learn
more about this function at [Ariba's Wiki](https://github.com/sanger-pathogens/ariba/wiki/Task:-getref).
1. __[ARG-ANNOT](http://en.mediterranee-infection.com/article.php?laref=283%26titre=arg-annot)__
Gupta SK, Padmanabhan BR, Diene SM, Lopez-Rojas R, Kempf M, Landraud L, Rolain J-M [ARG-ANNOT, a new bioinformatic tool to discover antibiotic resistance genes in bacterial genomes.](https://doi.org/10.1128/aac.01310-13) _Antimicrob. Agents Chemother_ 58, 212–220 (2014)
2. __[CARD](https://card.mcmaster.ca/)__
Alcock BP, Raphenya AR, Lau TTY, Tsang KK, Bouchard M, Edalatmand A, Huynh W, Nguyen A-L V, Cheng AA, Liu S, Min SY, Miroshnichenko A, Tran H-K, Werfalli RE, Nasir JA, Oloni M, Speicher DJ, Florescu A, Singh B, Faltyn M, Hernandez-Koutoucheva A, Sharma AN, Bordeleau E, Pawlowski AC, Zubyk HL, Dooley D, Griffiths E, Maguire F, Winsor GL, Beiko RG, Brinkman FSL, Hsiao WWL, Domselaar GV, McArthur AG [CARD 2020: antibiotic resistome surveillance with the comprehensive antibiotic resistance database.](https://doi.org/10.1093/nar/gkz935) _Nucleic acids research_ 48.D1, D517-D525 (2020)
3. __[EcOH](https://dx.doi.org/10.1099%2Fmgen.0.000064)__
Ingle DJ, Valcanis M, Kuzevski A, Tauschek M, Inouye M, Stinear T, Levine MM, Robins-Browne RM, Holt KE [In silico serotyping of E. coli from short read data identifies limited novel O-loci but extensive diversity of O:H serotype combinations within and between pathogenic lineages.](https://doi.org/10.1099/mgen.0.000064) _Microbial Genomics_, 2(7), e000064. (2016)
4. __[MEGARes](https://megares.meglab.org/)__
Lakin SM, Dean C, Noyes NR, Dettenwanger A, Ross AS, Doster E, Rovira P, Abdo Z, Jones KL, Ruiz J, Belk KE, Morley PS, Boucher C [MEGARes: an antimicrobial resistance database for high throughput sequencing.](https://doi.org/10.1093/nar/gkw1009) _Nucleic Acids Res._ 45, D574–D580 (2017)
5. __[MEGARes 2.0](https://megares.meglab.org/)__
Doster E, Lakin SM, Dean CJ, Wolfe C, Young JG, Boucher C, Belk KE, Noyes NR, Morley PS [MEGARes 2.0: a database for classification of antimicrobial drug, biocide and metal resistance determinants in metagenomic sequence data.](https://doi.org/10.1093/nar/gkz1010) _Nucleic Acids Research_, 48(D1), D561–D569. (2020)
6. __[NCBI Reference Gene Catalog](https://www.ncbi.nlm.nih.gov/bioproject/?term=PRJNA313047)__
Feldgarden M, Brover V, Haft DH, Prasad AB, Slotta DJ, Tolstoy I, Tyson GH, Zhao S, Hsu C-H, McDermott PF, Tadesse DA, Morales C, Simmons M, Tillman G, Wasilenko J, Folster JP, Klimke W [Validating the NCBI AMRFinder Tool and Resistance Gene Database Using Antimicrobial Resistance Genotype-Phenotype Correlations in a Collection of NARMS Isolates](https://doi.org/10.1128/AAC.00483-19). _Antimicrob. Agents Chemother._ (2019)
7. __[PlasmidFinder](https://cge.cbs.dtu.dk/services/PlasmidFinder/)__
Carattoli A, Zankari E, García-Fernández A, Larsen MV, Lund O, Villa L, Aarestrup FM, Hasman H [In silico detection and typing of plasmids using PlasmidFinder and plasmid multilocus sequence typing.](https://doi.org/10.1128/aac.02412-14) _Antimicrob. Agents Chemother._ 58, 3895–3903 (2014)
8. __[ResFinder](https://cge.cbs.dtu.dk//services/ResFinder/)__
Zankari E, Hasman H, Cosentino S, Vestergaard M, Rasmussen S, Lund O, Aarestrup FM, Larsen MV [Identification of acquired antimicrobial resistance genes.](https://doi.org/10.1093/jac/dks261) _J. Antimicrob. Chemother._ 67, 2640–2644 (2012)
9. __[SRST2](https://github.com/katholt/srst2)__
Inouye M, Dashnow H, Raven L-A, Schultz MB, Pope BJ, Tomita T, Zobel J, Holt KE [SRST2: Rapid genomic surveillance for public health and hospital microbiology labs.](https://doi.org/10.1186/s13073-014-0090-6) _Genome Med._ 6, 90 (2014)
10. __[VFDB](http://www.mgc.ac.cn/VFs/)__
Chen L, Zheng D, Liu B, Yang J, Jin Q [VFDB 2016: hierarchical and refined dataset for big data analysis--10 years on.](https://doi.org/10.1093/nar/gkv1239) _Nucleic Acids Res._ 44, D694–7 (2016)
11. __[VirulenceFinder](https://cge.cbs.dtu.dk/services/VirulenceFinder/)__
Joensen KG, Scheutz F, Lund O, Hasman H, Kaas RS, Nielsen EM, Aarestrup FM [Real-time whole-genome sequencing for routine typing, surveillance, and outbreak detection of verotoxigenic _Escherichia coli_.](https://doi.org/10.1128/jcm.03617-13) _J. Clin. Microbiol._ 52, 1501–1510 (2014)
### Minmer Datasets
1. __[Mash Refseq (release 88) Sketch](https://mash.readthedocs.io/en/latest/data.html)__
Ondov BD, Starrett GJ, Sappington A, Kostic A, Koren S, Buck CB, Phillippy AM [Mash Screen: high-throughput sequence containment estimation for genome discovery](https://doi.org/10.1186/s13059-019-1841-x) _Genome Biol_ 20, 232 (2019)
2. __[Sourmash Genbank LCA Signature](https://sourmash.readthedocs.io/en/latest/databases.html)__
Brown CT, Irber L [sourmash: a library for MinHash sketching of DNA](http://dx.doi.org/10.21105/joss.00027). _JOSS_ 1, 27 (2016)
### Everything Else
1. __[eggNOG 5.0 Database](http://eggnog.embl.de/)__
Huerta-Cepas J, Szklarczyk D, Heller D, Hernández-Plaza A, Forslund SK, Cook H, Mende DR, Letunic I, Rattei T, Jensen LJ, von Mering C, Bork P [eggNOG 5.0: a hierarchical, functionally and phylogenetically annotated orthology resource based on 5090 organisms and 2502 viruses.](https://doi.org/10.1093/nar/gky1085) _Nucleic Acids Res._ 47, D309–D314 (2019)
2. __[Genome Taxonomy Database](https://gtdb.ecogenomic.org/)__
Parks DH, Chuvochina M, Rinke C, Mussig AJ, Chaumeil P-A, Hugenholtz P [GTDB: an ongoing census of bacterial and archaeal diversity through a phylogenetically consistent, rank normalized and complete genome-based taxonomy](https://doi.org/10.1093/nar/gkab776) _Nucleic Acids Research_ gkab776 (2021)
3. __[MOB-suite Database](https://github.com/phac-nml/mob-suite)__
Robertson J, Bessonov K, Schonfeld J, Nash JHE. [Universal whole-sequence-based plasmid typing and its utility to prediction of host range and epidemiological surveillance.](https://doi.org/10.1099/mgen.0.000435) _Microbial Genomics_, 6(10)(2020)
4. __[NCBI RefSeq Database](https://www.ncbi.nlm.nih.gov/refseq/)__
O'Leary NA, Wright MW, Brister JR, Ciufo S, Haddad D, McVeigh R, Rajput B, Robbertse B, Smith-White B, Ako-Adjei D, Astashyn A, Badretdin A, Bao Y, Blinkova O0, Brover V, Chetvernin V, Choi J, Cox E, Ermolaeva O, Farrell CM, Goldfarb T, Gupta T, Haft D, Hatcher E, Hlavina W, Joardar VS, Kodali VK, Li W, Maglott D, Masterson P, McGarvey KM, Murphy MR, O'Neill K, Pujar S, Rangwala SH, Rausch D, Riddick LD, Schoch C, Shkeda A, Storz SS, Sun H, Thibaud-Nissen F, Tolstoy I, Tully RE, Vatsan AR, Wallin C, Webb D, Wu W, Landrum MJ, Kimchi A, Tatusova T, DiCuccio M, Kitts P, Murphy TD, Pruitt KD [Reference sequence (RefSeq) database at NCBI: current status, taxonomic expansion, and functional annotation.](https://doi.org/10.1093/nar/gkv1189) _Nucleic Acids Res._ 44, D733–45 (2016)
5. __[PubMLST.org](https://pubmlst.org/)__
Jolley KA, Bray JE, Maiden MCJ [Open-access bacterial population genomics: BIGSdb software, the PubMLST.org website and their applications.](http://dx.doi.org/10.12688/wellcomeopenres.14826.1) _Wellcome Open Res_ 3, 124 (2018)
6. __[SILVA rRNA Database](https://www.arb-silva.de/)__
Quast C, Pruesse E, Yilmaz P, Gerken J, Schweer T, Yarza P, Peplies J, Glöckner FO [The SILVA ribosomal RNA gene database project: improved data processing and web-based tools.](https://doi.org/10.1093/nar/gks1219) _Nucleic Acids Res._ 41, D590–6 (2013)
## Software Included In Bactopia
Below are 125 of software packages used (directly and indirectly) by
Bactopia. A link to the software page as well as the citation (if available)
have been included.
1. __[Abricate](https://github.com/tseemann/abricate)__
Mass screening of contigs for antimicrobial and virulence genes
Seemann T [Abricate: mass screening of contigs for antimicrobial and virulence genes](https://github.com/tseemann/abricate) (GitHub)
2. __[AgrVATE](https://github.com/VishnuRaghuram94/AgrVATE)__
Rapid identification of Staphylococcus aureus agr locus type and agr operon variants.
Raghuram V. [AgrVATE: Rapid identification of Staphylococcus aureus agr locus type and agr operon variants.](https://github.com/VishnuRaghuram94/AgrVATE) (GitHub)
3. __[AMRFinderPlus](https://github.com/ncbi/amr)__
Find acquired antimicrobial resistance genes and some point mutations in protein or assembled nucleotide sequences.
Feldgarden M, Brover V, Haft DH, Prasad AB, Slotta DJ, Tolstoy I, Tyson GH, Zhao S, Hsu C-H, McDermott PF, Tadesse DA, Morales C, Simmons M, Tillman G, Wasilenko J, Folster JP, Klimke W [Validating the NCBI AMRFinder Tool and Resistance Gene Database Using Antimicrobial Resistance Genotype-Phenotype Correlations in a Collection of NARMS Isolates](https://doi.org/10.1128/AAC.00483-19). _Antimicrob. Agents Chemother._ (2019)
4. __[any2fasta](https://github.com/tseemann/any2fasta)__
Convert various sequence formats to FASTA
Seemann T [any2fasta: Convert various sequence formats to FASTA](https://github.com/tseemann/any2fasta) (GitHub)
5. __[Aragorn](http://<IP_ADDRESS>/ARAGORN/Downloads/)__
Finds transfer RNA features (tRNA)
Laslett D, Canback B [ARAGORN, a program to detect tRNA genes and tmRNA genes in nucleotide sequences.](https://doi.org/10.1093/nar/gkh152) _Nucleic Acids Res_. 32(1):11-6 (2004)
6. __[Ariba](https://github.com/sanger-pathogens/ariba)__
Antimicrobial Resistance Identification By Assembly
Hunt M, Mather AE, Sánchez-Busó L, Page AJ, Parkhill J, Keane JA, Harris SR [ARIBA: rapid antimicrobial resistance genotyping directly from sequencing reads](http://dx.doi.org/10.1099/mgen.0.000131). _Microb Genom_ 3, e000131 (2017)
7. __[ART](https://www.niehs.nih.gov/research/resources/software/biostatistics/art/index.cfm)__
A set of simulation tools to generate synthetic next-generation sequencing reads
Huang W, Li L, Myers JR, Marth GT [ART: a next-generation sequencing read simulator.](http://dx.doi.org/10.1093/bioinformatics/btr708) _Bioinformatics_ 28, 593–594 (2012)
8. __[assembly-scan](https://github.com/rpetit3/assembly-scan)__
Generate basic stats for an assembly.
Petit III RA [assembly-scan: generate basic stats for an assembly](https://github.com/rpetit3/assembly-scan) (GitHub)
9. __[Bakta](https://github.com/oschwengers/bakta)__
Rapid & standardized annotation of bacterial genomes & plasmids
Schwengers O, Jelonek L, Dieckmann MA, Beyvers S, Blom J, Goesmann A [Bakta - rapid and standardized annotation of bacterial genomes via alignment-free sequence identification.](https://doi.org/10.1099/mgen.0.000685) _Microbial Genomics_ 7(11) (2021)
10. __[Barrnap](https://github.com/tseemann/barrnap)__
Bacterial ribosomal RNA predictor
Seemann T [Barrnap: Bacterial ribosomal RNA predictor](https://github.com/tseemann/barrnap) (GitHub)
11. __[BBTools](https://jgi.doe.gov/data-and-tools/bbtools/)__
BBTools is a suite of fast, multithreaded bioinformatics tools designed for analysis of DNA and RNA sequence data.
Bushnell B [BBMap short read aligner, and other bioinformatic tools.](http://sourceforge.net/projects/bbmap/) (Link)
12. __[BCFtools](https://github.com/samtools/bcftools)__
Utilities for variant calling and manipulating VCFs and BCFs.
Danecek P, Bonfield JK, Liddle J, Marshall J, Ohan V, Pollard MO, Whitwham A, Keane T, McCarthy SA, Davies RM, Li H [Twelve years of SAMtools and BCFtools](https://doi.org/10.1093/gigascience/giab008) _GigaScience_ Volume 10, Issue 2 (2021)
13. __[Bedtools](https://github.com/arq5x/bedtools2)__
A powerful toolset for genome arithmetic.
Quinlan AR, Hall IM [BEDTools: a flexible suite of utilities for comparing genomic features](http://dx.doi.org/10.1093/bioinformatics/btq033). _Bioinformatics_ 26, 841–842 (2010)
14. __[BLAST](https://blast.ncbi.nlm.nih.gov/Blast.cgi)__
Basic Local Alignment Search Tool
Camacho C, Coulouris G, Avagyan V, Ma N, Papadopoulos J, Bealer K, Madden TL [BLAST+: architecture and applications](http://dx.doi.org/10.1186/1471-2105-10-421). _BMC Bioinformatics_ 10, 421 (2009)
15. __[Bowtie2](https://github.com/BenLangmead/bowtie2)__
A fast and sensitive gapped read aligner
Langmead B, Salzberg SL [Fast gapped-read alignment with Bowtie 2.](http://dx.doi.org/10.1038/nmeth.1923) _Nat. Methods._ 9, 357–359 (2012)
16. __[BUSCO](https://gitlab.com/ezlab/busco)__
Assessing genome assembly and annotation completeness with Benchmarking Universal Single-Copy Orthologs (BUSCO)
Manni M, Berkeley MR, Seppey M, Simão FA, Zdobnov EM [BUSCO Update: Novel and Streamlined Workflows along with Broader and Deeper Phylogenetic Coverage for Scoring of Eukaryotic, Prokaryotic, and Viral Genomes.](https://doi.org/10.1093/molbev/msab199) _Molecular Biology and Evolution_ 38(10), 4647–4654. (2021)
17. __[BWA](https://github.com/lh3/bwa/)__
Burrow-Wheeler Aligner for short-read alignment
Li H [Aligning sequence reads, clone sequences and assembly contigs with BWA-MEM](http://arxiv.org/abs/1303.3997). _arXiv_ [q-bio.GN] (2013)
18. __[CD-HIT](https://github.com/weizhongli/cdhit)__
Accelerated for clustering the next-generation sequencing data
Li W, Godzik A [Cd-hit: a fast program for clustering and comparing large sets of protein or nucleotide sequences](http://dx.doi.org/10.1093/bioinformatics/btl158). _Bioinformatics_ 22, 1658–1659 (2006)
19. __[CD-HIT-EST](https://github.com/weizhongli/cdhit)__
Accelerated for clustering the next-generation sequencing data
Fu L, Niu B, Zhu Z, Wu S, Li W [CD-HIT: accelerated for clustering the next-generation sequencing data](http://dx.doi.org/10.1093/bioinformatics/bts565). _Bioinformatics_ 28, 3150–3152 (2012)
20. __[CheckM](https://github.com/Ecogenomics/CheckM)__
Assess the quality of microbial genomes recovered from isolates, single cells, and metagenomes
Parks DH, Imelfort M, Skennerton CT, Hugenholtz P, Tyson GW [CheckM: assessing the quality of microbial genomes recovered from isolates, single cells, and metagenomes.](http://dx.doi.org/10.1101/gr.186072.114) _Genome Res_ 25, 1043–1055 (2015)
21. __[ClonalFramML](https://github.com/xavierdidelot/ClonalFrameML)__
Efficient Inference of Recombination in Whole Bacterial Genomes
Didelot X, Wilson DJ [ClonalFrameML: Efficient Inference of Recombination in Whole Bacterial Genomes.](https://doi.org/10.1371/journal.pcbi.1004041) _PLoS Comput Biol_ 11(2) e1004041 (2015)
22. __[csvtk](https://bioinf.shenwei.me/csvtk/)__
A cross-platform, efficient and practical CSV/TSV toolkit in Golang
Shen, W [csvtk: A cross-platform, efficient and practical CSV/TSV toolkit in Golang.](https://github.com/shenwei356/csvtk/) (GitHub)
23. __[DIAMOND](https://github.com/bbuchfink/diamond)__
Accelerated BLAST compatible local sequence aligner.
Buchfink B, Xie C, Huson DH [Fast and sensitive protein alignment using DIAMOND.](http://dx.doi.org/10.1038/nmeth.3176) _Nat. Methods._ 12, 59–60 (2015)
24. __[ECTyper](https://github.com/phac-nml/ecoli_serotyping)__
In-silico prediction of _Escherichia coli_ serotype
Laing C, Bessonov K, Sung S, La Rose C [ECTyper - In silico prediction of _Escherichia coli_ serotype](https://github.com/phac-nml/ecoli_serotyping) (GitHub)
25. __[eggNOG-mapper](https://github.com/eggnogdb/eggnog-mapper)__
Fast genome-wide functional annotation through orthology assignment
Huerta-Cepas J, Forslund K, Coelho LP, Szklarczyk D, Jensen LJ, von Mering C, Bork P [Fast Genome-Wide Functional Annotation through Orthology Assignment by eggNOG-Mapper.](http://dx.doi.org/10.1093/molbev/msx148) _Mol. Biol. Evol._ 34, 2115–2122 (2017)
26. __[emmtyper](https://github.com/MDU-PHL/emmtyper)__
emm Automatic Isolate Labeller
Tan A, Seemann T, Lacey D, Davies M, Mcintyre L, Frost H, Williamson D, Gonçalves da Silva A [emmtyper - emm Automatic Isolate Labeller](https://github.com/MDU-PHL/emmtyper) (GitHub)
27. __[FastANI](https://github.com/ParBLiSS/FastANI)__
Fast Whole-Genome Similarity (ANI) Estimation
Jain C, Rodriguez-R LM, Phillippy AM, Konstantinidis KT, Aluru S [High throughput ANI analysis of 90K prokaryotic genomes reveals clear species boundaries.](http://dx.doi.org/10.1038/s41467-018-07641-9) _Nat. Commun._ 9, 5114 (2018)
28. __[FastQC](https://github.com/s-andrews/FastQC)__
A quality control analysis tool for high throughput sequencing data.
Andrews S [FastQC: a quality control tool for high throughput sequence data.](http://www.bioinformatics.babraham.ac.uk/projects/fastqc) (WebLink)
29. __[fastq-dl](https://github.com/rpetit3/fastq-dl)__
Download FASTQ files from SRA or ENA repositories.
Petit III RA [fastq-dl: Download FASTQ files from SRA or ENA repositories.](https://github.com/rpetit3/fastq-dl) (GitHub)
30. __[fastq-scan](https://github.com/rpetit3/fastq-scan)__
Output FASTQ summary statistics in JSON format
Petit III RA [fastq-scan: generate summary statistics of input FASTQ sequences.](https://github.com/rpetit3/fastq-scan) (GitHub)
31. __[FastTree](http://www.microbesonline.org/fasttree)__
Approximately-maximum-likelihood phylogenetic trees
Price MN, Dehal PS, Arkin AP [FastTree 2 – Approximately Maximum-Likelihood Trees for Large Alignments.](https://dx.doi.org/10.1371%2Fjournal.pone.0009490) _PLoS One_ 5, e9490 (2010)
32. __[FLASH](https://ccb.jhu.edu/software/FLASH/)__
A fast and accurate tool to merge paired-end reads.
Magoč T, Salzberg SL [FLASH: fast length adjustment of short reads to improve genome assemblies.](https://doi.org/10.1093/bioinformatics/btr507) _Bioinformatics_ 27.21 2957-2963 (2011)
33. __[Flye](https://github.com/fenderglass/Flye)__
De novo assembler for single molecule sequencing reads using repeat graphs
Kolmogorov M, Yuan J, Lin Y, Pevzner P [Assembly of Long Error-Prone Reads Using Repeat Graphs](https://doi.org/10.1038/s41587-019-0072-8) _Nature Biotechnology_ (2019)
34. __[freebayes](https://github.com/ekg/freebayes)__
Bayesian haplotype-based genetic polymorphism discovery and genotyping
Garrison E, Marth G [Haplotype-based variant detection from short-read sequencing.](https://arxiv.org/abs/1207.3907) arXiv preprint arXiv:1207.3907 [q-bio.GN] (2012)
35. __[GAMMA](https://github.com/rastanton/GAMMA)__
Gene Allele Mutation Microbial Assessment
Stanton RA, Vlachos N, Halpin AL [GAMMA: a tool for the rapid identification, classification, and annotation of translated gene matches from sequencing data.](https://doi.org/10.1093/bioinformatics/btab607) _Bioinformatics_ (2021)
36. __[GenoTyphi](https://github.com/katholt/genotyphi)__
Assign genotypes to Salmonella Typhi genomes based on Mykrobe results
Wong VK, Baker S, Connor TR, Pickard D, Page AJ, Dave J, Murphy N, Holliman R, Sefton A, Millar M, Dyson ZA, Dougan G, Holt KE, & International Typhoid Consortium. [An extended genotyping framework for Salmonella enterica serovar Typhi, the cause of human typhoid](https://doi.org/10.1038/ncomms12827) _Nature Communications_ 7, 12827. (2016)
37. __[GNU Parallel](https://www.gnu.org/software/parallel/)__
A shell tool for executing jobs in parallel
Tange O [GNU Parallel](https://doi.org/10.5281/zenodo.1146014) (2018)
38. __[GTDB-Tk](https://github.com/Ecogenomics/GTDBTk)__
A toolkit for assigning objective taxonomic classifications to bacterial and archaeal genomes
Chaumeil PA, Mussig AJ, Hugenholtz P, Parks DH [GTDB-Tk: a toolkit to classify genomes with the Genome Taxonomy Database.](https://doi.org/10.1093/bioinformatics/btz848) _Bioinformatics_ (2019)
39. __[Gubbins](https://github.com/nickjcroucher/gubbins)__
Rapid phylogenetic analysis of large samples of recombinant bacterial whole genome sequences
Croucher NJ, Page AJ, Connor TR, Delaney AJ, Keane JA, Bentley SD, Parkhill J, Harris SR [Rapid phylogenetic analysis of large samples of recombinant bacterial whole genome sequences using Gubbins.](https://doi.org/10.1093/nar/gku1196) _Nucleic Acids Research_ 43(3), e15. (2015)
40. __[hicap](https://github.com/scwatts/hicap)__
in silico typing of the _H. influenzae_ cap locus
Watts SC, Holt KE [hicap: in silico serotyping of the Haemophilus influenzae capsule locus.](https://doi.org/10.1128/JCM.00190-19) _Journal of Clinical Microbiology_ JCM.00190-19 (2019)
41. __[HMMER](http://hmmer.org/)__
Biosequence analysis using profile hidden Markov models
Eddy SR [Accelerated Profile HMM Searches.](https://doi.org/10.1371/journal.pcbi.1002195) _PLoS Comput. Biol._ 7, e1002195 (2011)
42. __[HpsuisSero](https://github.com/jimmyliu1326/HpsuisSero)__
Rapid _Haemophilus parasuis_ serotyping
Lui J [HpsuisSero: Rapid _Haemophilus parasuis_ serotyping](https://github.com/jimmyliu1326/HpsuisSero) (GitHub)
43. __[Infernal](http://eddylab.org/infernal/)__
Searches DNA sequence databases for RNA structure and sequence similarities
Nawrocki EP, Eddy SR [Infernal 1.1: 100-fold faster RNA homology searches.](https://doi.org/10.1093/bioinformatics/btt509) _Bioinformatics_ 29(22), 2933-2935 (2013)
44. __[IQ-TREE](https://github.com/Cibiv/IQ-TREE)__
Efficient phylogenomic software by maximum likelihood
Nguyen L-T, Schmidt HA, von Haeseler A, Minh BQ [IQ-TREE: A fast and effective stochastic algorithm for estimating maximum likelihood phylogenies.](https://doi.org/10.1093/molbev/msu300) _Mol. Biol. Evol._ 32:268-274 (2015)
45. __[ModelFinder](https://github.com/Cibiv/IQ-TREE)__
Used for automatic model selection
Kalyaanamoorthy S, Minh BQ, Wong TKF, von Haeseler A, Jermiin LS [ModelFinder - Fast model selection for accurate phylogenetic estimates.](https://doi.org/10.1038/nmeth.4285) _Nat. Methods_ 14:587-589 (2017)
46. __[UFBoot2](https://github.com/Cibiv/IQ-TREE)__
Used to conduct ultrafast bootstrapping
Hoang DT, Chernomor O, von Haeseler A, Minh BQ, Vinh LS [UFBoot2: Improving the ultrafast bootstrap approximation.](https://doi.org/10.1093/molbev/msx281) _Mol. Biol. Evol._ 35:518–522 (2018)
47. __[ISMapper](https://github.com/jhawkey/IS_mapper)__
IS mapping software
Hawkey J, Hamidian M, Wick RR, Edwards DJ, Billman-Jacobe H, Hall RM, Holt KE [ISMapper: identifying transposase insertion sites in bacterial genomes from short read sequence data](http://dx.doi.org/10.1186/s12864-015-1860-2). _BMC Genomics_ 16, 667 (2015)
48. __[Kaptive](https://github.com/katholt/Kaptive)__
Surface polysaccharide loci for _Klebsiella pneumoniae_ species complex and _Acinetobacter baumannii_ genomes
Wyres KL, Wick RR, Gorrie C, Jenney A, Follador R, Thomson NR, Holt KE [Identification of Klebsiella capsule synthesis loci from whole genome data.](https://doi.org/10.1099/mgen.0.000102) _Microbial genomics_ 2(12) (2016)
49. __[Kleborate](https://github.com/katholt/Kleborate)__
Genotyping tool for _Klebsiella pneumoniae_ and its related species complex
Lam MMC, Wick RR, Watts, SC, Cerdeira LT, Wyres KL, Holt KE [A genomic surveillance framework and genotyping tool for Klebsiella pneumoniae and its related species complex.](https://doi.org/10.1038/s41467-021-24448-3) _Nat Commun_ 12, 4188 (2021)
50. __[KMC](https://github.com/refresh-bio/KMC)__
Fast and frugal disk based k-mer counter
Deorowicz S, Kokot M, Grabowski Sz, Debudaj-Grabysz A [KMC 2: Fast and resource-frugal k-mer counting](https://doi.org/10.1093/bioinformatics/btv022) _Bioinformatics_ 31(10):1569–1576 (2015)
51. __[Kraken2](https://github.com/DerrickWood/kraken2)__
The second version of the Kraken taxonomic sequence classification system
Wood DE, Lu J, Langmead B [Improved metagenomic analysis with Kraken 2.](https://doi.org/10.1186/s13059-019-1891-0) *Genome Biology*, 20(1), 257. (2019)
52. __[legsta](https://github.com/tseemann/legsta)__
In silico Legionella pneumophila Sequence Based Typing
Seemann T [legsta: In silico Legionella pneumophila Sequence Based Typing](https://github.com/tseemann/legsta) (GitHub)
53. __[Lighter](https://github.com/mourisl/Lighter)__
Fast and memory-efficient sequencing error corrector
Song L, Florea L, Langmead B [Lighter: Fast and Memory-efficient Sequencing Error Correction without Counting](https://doi.org/10.1186/s13059-014-0509-9). _Genome Biol._ 15(11):509 (2014)
54. __[LisSero](https://github.com/MDU-PHL/LisSero)__
_In silico_ serotype prediction for _Listeria monocytogenes_
Kwong J, Zhang J, Seeman T, Horan, K, Gonçalves da Silva A [LisSero - _In silico_ serotype prediction for _Listeria monocytogenes_](https://github.com/MDU-PHL/LisSero) (GitHub)
55. __[MAFFT](https://mafft.cbrc.jp/alignment/software/)__
Multiple alignment program for amino acid or nucleotide sequences
Katoh K, Standley DM [MAFFT multiple sequence alignment software version 7: improvements in performance and usability.](https://doi.org/10.1093/molbev/mst010) _Mol. Biol. Evol._ 30, 772–780 (2013)
56. __[Mash](https://github.com/marbl/Mash)__
Fast genome and metagenome distance estimation using MinHash
Ondov BD, Treangen TJ, Melsted P, Mallonee AB, Bergman NH, Koren S, Phillippy AM [Mash: fast genome and metagenome distance estimation using MinHash](http://dx.doi.org/10.1186/s13059-016-0997-x). _Genome Biol_ 17, 132 (2016)
57. __[Mash](https://github.com/marbl/Mash)__
High-throughput sequence containment estimation
Ondov BD, Starrett GJ, Sappington A, Kostic A, Koren S, Buck CB, Phillippy AM [Mash Screen: high-throughput sequence containment estimation for genome discovery](https://doi.org/10.1186/s13059-019-1841-x) _Genome Biol_ 20, 232 (2019)
58. __[Mashtree](https://github.com/lskatz/mashtree)__
Create a tree using Mash distances
Katz LS, Griswold T, Morrison S, Caravas J, Zhang S, den Bakker HC, Deng X, Carleton HA [Mashtree: a rapid comparison of whole genome sequence files.](https://doi.org/10.21105/joss.01762) _Journal of Open Source Software_, 4(44), 1762 (2019)
59. __[maskrc-svg](https://github.com/kwongj/maskrc-svg)__
Masks recombination as detected by ClonalFrameML or Gubbins
Kwong J [maskrc-svg - Masks recombination as detected by ClonalFrameML or Gubbins and draws an SVG.](https://github.com/kwongj/maskrc-svg) (GitHub)
60. __[McCortex](https://github.com/mcveanlab/mccortex)__
De novo genome assembly and multisample variant calling
Turner I, Garimella KV, Iqbal Z, McVean G [Integrating long-range connectivity information into de Bruijn graphs.](http://dx.doi.org/10.1093/bioinformatics/bty157) _Bioinformatics_ 34, 2556–2565 (2018)
61. __[mcroni](https://github.com/liampshaw/mcroni)__
Scripts for finding and processing promoter variants upstream of mcr-1
Shaw L [mcroni: Scripts for finding and processing promoter variants upstream of mcr-1](https://github.com/liampshaw/mcroni) (GitHub)
62. __[Medaka](None)__
Sequence correction provided by ONT Research
ONT Research [Medaka: Sequence correction provided by ONT Research](https://github.com/nanoporetech/medaka) (GitHub)
63. __[meningotype](https://github.com/MDU-PHL/meningotype)__
In silico serotyping, finetyping and Bexsero antigen sequence typing of _Neisseria meningitidis_
Kwong JC, Gonçalves da Silva A, Stinear TP, Howden BP, & Seemann T [meningotype: in silico typing for _Neisseria meningitidis_.](https://github.com/MDU-PHL/meningotype) (GitHub)
64. __[MEGAHIT](https://github.com/voutcn/megahit)__
Ultra-fast and memory-efficient (meta-)genome assembler
Li D, Liu C-M, Luo R, Sadakane K, Lam T-W [MEGAHIT: an ultra-fast single-node solution for large and complex metagenomics assembly via succinct de Bruijn graph.](https://doi.org/10.1093/bioinformatics/btv033) _Bioinformatics_ 31.10 1674-1676 (2015)
65. __[mlst](https://github.com/tseemann/mlst)__
Scan contig files against PubMLST typing schemes
Seemann T [mlst: scan contig files against PubMLST typing schemes](https://github.com/tseemann/mlst) (GitHub)
66. __[MinCED](https://github.com/ctSkennerton/minced)__
Mining CRISPRs in Environmental Datasets
Skennerton C [MinCED: Mining CRISPRs in Environmental Datasets](https://github.com/ctSkennerton/minced) (GitHub)
67. __[Miniasm](https://github.com/lh3/miniasm)__
Ultrafast de novo assembly for long noisy reads (though having no consensus step)
Li H [Miniasm: Ultrafast de novo assembly for long noisy reads](https://github.com/lh3/miniasm) (GitHub)
68. __[Minimap2](https://github.com/lh3/minimap2)__
A versatile pairwise aligner for genomic and spliced nucleotide sequences
Li H [Minimap2: pairwise alignment for nucleotide sequences.](https://doi.org/10.1093/bioinformatics/bty191) _Bioinformatics_ 34:3094-3100 (2018)
69. __[MOB-suite](https://github.com/phac-nml/mob-suite)__
Software tools for clustering, reconstruction and typing of plasmids from draft assemblies
Robertson J, Nash JHE [MOB-suite: software tools for clustering, reconstruction and typing of plasmids from draft assemblies.](https://doi.org/10.1099/mgen.0.000206) _Microbial Genomics_ 4(8). (2018)
70. __[Mykrobe](https://github.com/Mykrobe-tools/mykrobe)__
Antibiotic resistance prediction in minutes
Hunt M, Bradley P, Lapierre SG, Heys S, Thomsit M, Hall MB, Malone KM, Wintringer P, Walker TM, Cirillo DM, Comas I, Farhat MR, Fowler P, Gardy J, Ismail N, Kohl TA, Mathys V, Merker M, Niemann S, Omar SV, Sintchenko V, Smith G, Supply P, Tahseen S, Wilcox M, Arandjelovic I, Peto TEA, Crook, DW, Iqbal Z [Antibiotic resistance prediction for Mycobacterium tuberculosis from genome sequence data with Mykrobe](https://doi.org/10.12688/wellcomeopenres.15603.1) _Wellcome Open Research_ 4, 191. (2019)
71. __[NanoPlot](https://github.com/wdecoster/NanoPlot)__
Plotting scripts for long read sequencing data
De Coster W, D’Hert S, Schultz DT, Cruts M, Van Broeckhoven C [NanoPack: visualizing and processing long-read sequencing data](https://doi.org/10.1093/bioinformatics/bty149) _Bioinformatics_ Volume 34, Issue 15 (2018)
72. __[Nanoq](https://github.com/esteinig/nanoq)__
Minimal but speedy quality control for nanopore reads in Rust
Steinig E [Nanoq: Minimal but speedy quality control for nanopore reads in Rust](https://github.com/esteinig/nanoq) (GitHub)
73. __[ncbi-genome-download](https://github.com/kblin/ncbi-genome-download)__
Scripts to download genomes from the NCBI FTP servers
Blin K [ncbi-genome-download: Scripts to download genomes from the NCBI FTP servers](https://github.com/kblin/ncbi-genome-download) (GitHub)
74. __[Nextflow](https://github.com/nextflow-io/nextflow)__
A DSL for data-driven computational pipelines.
Di Tommaso P, Chatzou M, Floden EW, Barja PP, Palumbo E, Notredame C [Nextflow enables reproducible computational workflows.](https://www.nature.com/articles/nbt.3820.pdf?origin=ppub) _Nat. Biotechnol._ 35, 316–319 (2017)
75. __[ngmaster](https://github.com/MDU-PHL/ngmaster)__
_In silico_ multi-antigen sequence typing for _Neisseria gonorrhoeae_ (NG-MAST)
Kwong J, Gonçalves da Silva A, Schultz M, Seeman T [ngmaster - _In silico_ multi-antigen sequence typing for _Neisseria gonorrhoeae_ (NG-MAST)](https://github.com/MDU-PHL/ngmaster) (GitHub)
76. __[nhmmer](http://hmmer.org/)__
DNA homology search with profile HMMs.
Wheeler TJ, Eddy SR [nhmmer: DNA homology search with profile HMMs.](https://doi.org/10.1093/bioinformatics/btt403) _Bioinformatics_ 29, 2487–2489 (2013)
77. __[Panaroo](https://github.com/gtonkinhill/panaroo)__
An updated pipeline for pangenome investigation
Tonkin-Hill G, MacAlasdair N, Ruis C, Weimann A, Horesh G, Lees JA, Gladstone RA, Lo S, Beaudoin C, Floto RA, Frost SDW, Corander J, Bentley SD, Parkhill J [Producing polished prokaryotic pangenomes with the Panaroo pipeline.](https://doi.org/10.1186/s13059-020-02090-4) _Genome Biology_ 21(1), 180. (2020)
78. __[pasty](https://github.com/rpetit3/pasty)__
in silico serogrouping of _Pseudomonas aeruginosa_ isolates
Petit III RA [pasty: in silico serogrouping of _Pseudomonas aeruginosa_ isolates](https://github.com/rpetit3/pasty) (GitHub)
79. __[pbptyper](https://github.com/rpetit3/pbptyper)__
Penicillin Binding Protein (PBP) typer for _Streptococcus pneumoniae_ assemblies
Petit III RA [pbptyper: In silico Penicillin Binding Protein (PBP) typer for _Streptococcus pneumoniae_ assemblies](https://github.com/rpetit3/pbptyper) (GitHub)
80. __[phyloFlash](https://github.com/HRGV/phyloFlash)__
A pipeline to rapidly reconstruct the SSU rRNAs and explore phylogenetic composition of an illumina (meta)genomic dataset.
Gruber-Vodicka HR, Seah BKB, Pruesse E [phyloFlash: Rapid Small-Subunit rRNA Profiling and Targeted Assembly from Metagenomes](https://doi.org/10.1128/mSystems.00920-20) _mSystems_ 5 (2020)
81. __[Pigz](https://zlib.net/pigz/)__
A parallel implementation of gzip for modern multi-processor, multi-core machines.
Adler M. [pigz: A parallel implementation of gzip for modern multi-processor, multi-core machines.](https://zlib.net/pigz/) _Jet Propulsion Laboratory_ (2015)
82. __[Pilon](https://github.com/broadinstitute/pilon/)__
An automated genome assembly improvement and variant detection tool
Walker BJ, Abeel T, Shea T, Priest M, Abouelliel A, Sakthikumar S, Cuomo CA, Zeng Q, Wortman J, Young SK, Earl AM [Pilon: an integrated tool for comprehensive microbial variant detection and genome assembly improvement.](https://doi.org/10.1371/journal.pone.0112963) _PloS one_ 9.11 e112963 (2014)
83. __[PIRATE](http://github.com/SionBayliss/PIRATE)__
A toolbox for pangenome analysis and threshold evaluation.
Bayliss SC, Thorpe HA, Coyle NM, Sheppard SK, Feil EJ [PIRATE: A fast and scalable pangenomics toolbox for clustering diverged orthologues in bacteria.](https://doi.org/10.1093/gigascience/giz119) _Gigascience_ 8 (2019)
84. __[PlasmidFinder](https://bitbucket.org/genomicepidemiology/plasmidfinder)__
Identifies plasmids in total or partial sequenced isolates of bacteria
Carattoli A, Zankari E, García-Fernández A, Voldby Larsen M, Lund O, Villa L, Møller Aarestrup F, Hasman H [In silico detection and typing of plasmids using PlasmidFinder and plasmid multilocus sequence typing.](https://doi.org/10.1128/AAC.02412-14) _Antimicrobial Agents and Chemotherapy_ 58(7), 3895–3903. (2014)
85. __[Porechop](https://github.com/rrwick/Porechop)__
adapter trimmer for Oxford Nanopore reads
Wick RR, Judd LM, Gorrie CL, Holt KE. [Completing bacterial genome assemblies with multiplex MinION sequencing.](https://doi.org/10.1099/mgen.0.000132) _Microb Genom._ 3(10):e000132 (2017)
86. __[pplacer](https://github.com/matsen/pplacer)__
Phylogenetic placement and downstream analysis
Matsen FA, Kodner RB, Armbrust EV [pplacer: linear time maximum-likelihood and Bayesian phylogenetic placement of sequences onto a fixed reference tree.](https://doi.org/10.1186/1471-2105-11-538) _BMC Bioinformatics_ 11, 538 (2010)
87. __[Prodigal](https://github.com/hyattpd/Prodigal)__
Fast, reliable protein-coding gene prediction for prokaryotic genomes.
Hyatt D, Chen G-L, LoCascio PF, Land ML, Larimer FW, Hauser LJ [Prodigal: prokaryotic gene recognition and translation initiation site identification.](https://doi.org/10.1186/1471-2105-11-119) _BMC Bioinformatics_ 11.1 119 (2010)
88. __[Prokka](https://github.com/tseemann/prokka)__
Rapid prokaryotic genome annotation
Seemann T [Prokka: rapid prokaryotic genome annotation](http://dx.doi.org/10.1093/bioinformatics/btu153) _Bioinformatics_ 30, 2068–2069 (2014)
89. __[QUAST](http://quast.sourceforge.net/)__
Quality Assessment Tool for Genome
Gurevich A, Saveliev V, Vyahhi N, Tesler G [QUAST: quality assessment tool for genome assemblies.](http://dx.doi.org/10.1093/bioinformatics/btt086) _Bioinformatics_ 29, 1072–1075 (2013)
90. __[Racon](https://github.com/lbcb-sci/racon)__
Ultrafast consensus module for raw de novo genome assembly of long uncorrected reads
Vaser R, Sović I, Nagarajan N, Šikić M [Fast and accurate de novo genome assembly from long uncorrected reads.](http://dx.doi.org/10.1101/gr.214270.116) _Genome Res_ 27, 737–746 (2017)
91. __[Rasusa](https://github.com/mbhall88/rasusa)__
Randomly subsample sequencing reads to a specified coverage
Hall MB [Rasusa: Randomly subsample sequencing reads to a specified coverage.](https://doi.org/10.5281/zenodo.3731394) (2019).
92. __[Raven](https://github.com/lbcb-sci/raven)__
De novo genome assembler for long uncorrected reads
Vaser R, Šikić M [Time- and memory-efficient genome assembly with Raven.](https://doi.org/10.1038/s43588-021-00073-4) _Nat Comput Sci_ 1, 332–336 (2021)
93. __[Resistance Gene Identifier (RGI)](https://github.com/arpcard/rgi)__
Software to predict resistomes from protein or nucleotide data, based on homology and SNP models.
Alcock BP, Raphenya AR, Lau TTY, Tsang KK, Bouchard M, Edalatmand A, Huynh W, Nguyen A-L V, Cheng AA, Liu S, Min SY, Miroshnichenko A, Tran H-K, Werfalli RE, Nasir JA, Oloni M, Speicher DJ, Florescu A, Singh B, Faltyn M, Hernandez-Koutoucheva A, Sharma AN, Bordeleau E, Pawlowski AC, Zubyk HL, Dooley D, Griffiths E, Maguire F, Winsor GL, Beiko RG, Brinkman FSL, Hsiao WWL, Domselaar GV, McArthur AG [CARD 2020: antibiotic resistome surveillance with the comprehensive antibiotic resistance database.](https://doi.org/10.1093/nar/gkz935) _Nucleic acids research_ 48.D1, D517-D525 (2020)
94. __[RNAmmer](http://www.cbs.dtu.dk/services/RNAmmer/)__
Consistent and rapid annotation of ribosomal RNA genes
Lagesen K, Hallin P, Rødland EA, Stærfeldt H-H, Rognes T, Ussery DW [RNAmmer: consistent annotation of rRNA genes in genomic sequences.](https://dx.doi.org/10.1093%2Fnar%2Fgkm160) _Nucleic Acids Res_ 35.9: 3100-3108 (2007)
95. __[Roary](https://github.com/sanger-pathogens/Roary)__
Rapid large-scale prokaryote pan genome analysis
Page AJ, Cummins CA, Hunt M, Wong VK, Reuter S, Holden MTG, Fookes M, Falush D, Keane JA, Parkhill J [Roary: rapid large-scale prokaryote pan genome analysis.](https://doi.org/10.1093/bioinformatics/btv421) _Bioinformatics_ 31, 3691–3693 (2015)
96. __[samclip](https://github.com/tseemann/samclip)__
Filter SAM file for soft and hard clipped alignments
Seemann T [Samclip: Filter SAM file for soft and hard clipped alignments](https://github.com/tseemann/samclip) (GitHub)
97. __[Samtools](https://github.com/samtools/samtools)__
Tools for manipulating next-generation sequencing data
Li H, Handsaker B, Wysoker A, Fennell T, Ruan J, Homer N, Marth G, Abecasis G, Durbin R [The Sequence Alignment/Map format and SAMtools](http://dx.doi.org/10.1093/bioinformatics/btp352). _Bioinformatics_ 25, 2078–2079 (2009)
98. __[Scoary](https://github.com/AdmiralenOla/Scoary)__
Pan-genome wide association studies
Brynildsrud O, Bohlin J, Scheffer L, Eldholm V [Rapid scoring of genes in microbial pan-genome-wide association studies with Scoary.](https://doi.org/10.1186/s13059-016-1108-8) _Genome Biol._ 17:238 (2016)
99. __[SeqSero2](https://github.com/denglab/SeqSero2)__
Salmonella serotype prediction from genome sequencing data
Zhang S, Den-Bakker HC, Li S, Dinsmore BA, Lane C, Lauer AC, Fields PI, Deng X. [SeqSero2: rapid and improved Salmonella serotype determination using whole genome sequencing data.](https://doi.org/10.1128/AEM.01746-19) _Appl Environ Microbiology_ 85(23):e01746-19 (2019)
100. __[Seqtk](https://github.com/lh3/seqtk)__
A fast and lightweight tool for processing sequences in the FASTA or FASTQ format.
Li H [Toolkit for processing sequences in FASTA/Q formats](https://github.com/lh3/seqtk) (GitHub)
101. __[Seroba](https://github.com/sanger-pathogens/seroba)__
k-mer based pipeline to identify the serotype of _Streptococcus pneumoniae_ from Illumina NGS reads
Epping L, van Tonder AJ, Gladstone RA, The Global Pneumococcal Sequencing Consortium, Bentley SD, Page AJ, Keane JA [SeroBA: rapid high-throughput serotyping of Streptococcus pneumoniae from whole genome sequence data.](https://doi.org/10.1099/mgen.0.000186) _Microbial Genomics_, 4(7) (2018)
102. __[ShigaTyper](https://github.com/CFSAN-Biostatistics/shigatyper)__
Shigella serotype from Illumina or Oxford Nanopore reads
Wu Y, Lau HK, Lee T, Lau DK, Payne J [In Silico Serotyping Based on Whole-Genome Sequencing Improves the Accuracy of Shigella Identification.](https://doi.org/10.1128/AEM.00165-19) *Applied and Environmental Microbiology*, 85(7). (2019)
103. __[ShigEiFinder](https://github.com/LanLab/ShigEiFinder)__
Cluster informed Shigella and EIEC serotyping tool from Illumina reads and assemblies
Zhang X, Payne M, Nguyen T, Kaur S, Lan R [Cluster-specific gene markers enhance Shigella and enteroinvasive Escherichia coli in silico serotyping.](https://doi.org/10.1099/mgen.0.000704) Microbial Genomics, 7(12). (2021)
104. __[Shovill](https://github.com/tseemann/shovill)__
Faster assembly of Illumina reads
Seemann T [Shovill: De novo assembly pipeline for Illumina paired reads](https://github.com/tseemann/shovill) (GitHub)
105. __[SignalP](http://www.cbs.dtu.dk/services/SignalP-4.0/)__
SISTR (Salmonella In Silico Typing Resource) command-line tool
Petersen TN, Brunak S, von Heijne G, Nielsen H [SignalP 4.0: discriminating signal peptides from transmembrane regions.](https://doi.org/10.1038/nmeth.1701) _Nature methods_ 8.10: 785 (2011)
106. __[SISTR](https://github.com/phac-nml/sistr_cmd)__
Finds signal peptide features in CDS
Yoshida CE, Kruczkiewicz P, Laing CR, Lingohr EJ, Gannon VPJ, Nash JHE, Taboada EN [The Salmonella In Silico Typing Resource (SISTR): An Open Web-Accessible Tool for Rapidly Typing and Subtyping Draft Salmonella Genome Assemblies.](https://doi.org/10.1371/journal.pone.0147101) _PloS One_, 11(1), e0147101. (2016)
107. __[SKESA](https://github.com/ncbi/SKESA)__
Strategic Kmer Extension for Scrupulous Assemblies
Souvorov A, Agarwala R, Lipman DJ [SKESA: strategic k-mer extension for scrupulous assemblies.](https://doi.org/10.1186/s13059-018-1540-z) _Genome Biology_ 19:153 (2018)
108. __[Snippy](https://github.com/tseemann/snippy)__
Rapid haploid variant calling and core genome alignment
Seemann T [Snippy: fast bacterial variant calling from NGS reads](https://github.com/tseemann/snippy) (GitHub)
109. __[SnpEff](http://snpeff.sourceforge.net/)__
Genomic variant annotations and functional effect prediction toolbox.
Cingolani P, Platts A, Wang LL, Coon M, Nguyen T, Wang L, Land SJ, Lu X, Douglas M [A program for annotating and predicting the effects of single nucleotide polymorphisms, SnpEff: SNPs in the genome of Drosophila melanogaster strain w1118; iso-2; iso-3.](https://doi.org/10.4161/fly.19695) _Fly_ 6(2), 80-92 (2012)
110. __[snp-dists](https://github.com/tseemann/snp-dists)__
Pairwise SNP distance matrix from a FASTA sequence alignment
Seemann T [snp-dists - Pairwise SNP distance matrix from a FASTA sequence alignment.](https://github.com/tseemann/snp-dists) (GitHub)
111. __[SNP-sites](https://github.com/sanger-pathogens/snp-sites)__
Rapidly extracts SNPs from a multi-FASTA alignment.
Page AJ, Taylor B, Delaney AJ, Soares J, Seemann T, Keane JA, Harris SR [SNP-sites: rapid efficient extraction of SNPs from multi-FASTA alignments.](https://dx.doi.org/10.1099%2Fmgen.0.000056) _Microbial Genomics_ 2.4 (2016)
112. __[Sourmash](https://github.com/dib-lab/sourmash)__
Compute and compare MinHash signatures for DNA data sets.
Brown CT, Irber L [sourmash: a library for MinHash sketching of DNA](http://dx.doi.org/10.21105/joss.00027). _JOSS_ 1, 27 (2016)
113. __[SPAdes](https://github.com/ablab/spades)__
An assembly toolkit containing various assembly pipelines.
Bankevich A, Nurk S, Antipov D, Gurevich AA, Dvorkin M, Kulikov AS, Lesin VM, Nikolenko SI, Pham S, Prjibelski AD, Pyshkin AV, Sirotkin AV, Vyahhi N, Tesler G, Alekseyev MA, Pevzner PA [SPAdes: a new genome assembly algorithm and its applications to single-cell sequencing.](https://doi.org/10.1089/cmb.2012.0021) _Journal of computational biology_ 19.5 455-477 (2012)
114. __[spaTyper](https://github.com/HCGB-IGTP/spaTyper)__
Computational method for finding spa types.
Sanchez-Herrero JF, and Sullivan M [spaTyper: Staphylococcal protein A (spa) characterization pipeline](http://doi.org/10.5281/zenodo.4063625). Zenodo. (2020)
115. __[spaTyper Database](https://cge.cbs.dtu.dk/services/spatyper/)__
Database used by spaTyper
Harmsen D, Claus H, Witte W, Rothgänger J, Claus H, Turnwald D, and Vogel U [Typing of methicillin-resistant _Staphylococcus aureus_ in a university hospital setting using a novel software for spa-repeat determination and database management.](https://doi.org/10.1128/jcm.41.12.5442-5448.2003) _J. Clin. Microbiol._ 41:5442-5448 (2003)
116. __[SsuisSero](https://github.com/jimmyliu1326/SsuisSero)__
Rapid _Streptococcus suis_ serotyping
Lui J [SsuisSero: Rapid _Streptococcus suis_ serotyping](https://github.com/jimmyliu1326/SsuisSero) (GitHub)
117. __[staphopia-sccmec](https://github.com/staphopia/staphopia-sccmec)__
A standalone version of Staphopia's SCCmec typing method.
Petit III RA, Read TD [_Staphylococcus aureus_ viewed from the perspective of 40,000+ genomes.](http://dx.doi.org/10.7717/peerj.5261) _PeerJ_ 6, e5261 (2018)
118. __[TBProfiler](https://github.com/jodyphelan/TBProfiler)__
Profiling tool for _Mycobacterium tuberculosis_ to detect resistance and strain type
Phelan JE, O’Sullivan DM, Machado D, Ramos J, Oppong YEA, Campino S, O’Grady J, McNerney R, Hibberd ML, Viveiros M, Huggett JF, Clark TG [Integrating informatics tools and portable sequencing technology for rapid detection of resistance to anti-tuberculous drugs.](https://doi.org/10.1186/s13073-019-0650-x) _Genome Med_ 11, 41 (2019)
119. __[Trimmomatic](http://www.usadellab.org/cms/index.php?page=trimmomatic)__
A flexible read trimming tool for Illumina NGS data
Bolger AM, Lohse M, Usadel B [Trimmomatic: a flexible trimmer for Illumina sequence data.](https://doi.org/10.1093/bioinformatics/btu170) _Bioinformatics_ 30.15 2114-2120 (2014)
120. __[Unicycler](https://github.com/rrwick/Unicycler)__
Hybrid assembly pipeline for bacterial genomes
Wick RR, Judd LM, Gorrie CL, Holt KE [Unicycler: Resolving bacterial genome assemblies from short and long sequencing reads.](http://dx.doi.org/10.1371/journal.pcbi.1005595) _PLoS Comput. Biol._ 13, e1005595 (2017)
121. __[VCF-Annotator](https://github.com/rpetit3/vcf-annotator)__
Add biological annotations to variants in a VCF file.
Petit III RA [VCF-Annotator: Add biological annotations to variants in a VCF file.](https://github.com/rpetit3/vcf-annotator) (GitHub)
122. __[Vcflib](https://github.com/vcflib/vcflib)__
a simple C++ library for parsing and manipulating VCF files
Garrison E [Vcflib: A C++ library for parsing and manipulating VCF files](https://github.com/vcflib/vcflib) (GitHub)
123. __[Velvet](https://github.com/dzerbino/velvet)__
Short read de novo assembler using de Bruijn graphs
Zerbino DR, Birney E [Velvet: algorithms for de novo short read assembly using de Bruijn graphs.](http://www.genome.org/cgi/doi/10.1101/gr.074492.107) _Genome research_ 18.5 821-829 (2008)
124. __[VSEARCH](https://github.com/torognes/vsearch)__
Versatile open-source tool for metagenomics
Rognes T, Flouri T, Nichols B, Quince C, Mahé F [VSEARCH: a versatile open source tool for metagenomics.](https://doi.org/10.7717/peerj.2584) _PeerJ_ 4, e2584 (2016)
125. __[vt](https://github.com/atks/vt)__
A tool set for short variant discovery in genetic sequence data.
Tan A, Abecasis GR, Kang HM [Unified representation of genetic variants.](https://doi.org/10.1093/bioinformatics/btv112) _Bioinformatics_ 31(13), 2202-2204 (2015)
|
2024-11-18T18:03:44.917603+00:00 | 2020-08-25T04:54:57 | 8d95d4272420f88a60d471759dcdc1fe35897226 | 4 | {
"blob_id": "8d95d4272420f88a60d471759dcdc1fe35897226",
"branch_name": "refs/heads/master",
"committer_date": "2020-08-25T04:54:57",
"content_id": "7f1b967f5b23ac804b5ff8f4bfd7a5cd1a633e25",
"detected_licenses": [
"MIT"
],
"directory_id": "e784f5dac744333897e67e3551b41c6af294a0d5",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": null,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 3336,
"license": "MIT",
"license_type": "permissive",
"path": "/create-first-app/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:44",
"repo_name": "enesusanmaz/azure-search-dotnet-samples",
"revision_date": "2020-08-25T04:54:57",
"revision_id": "42d75a0ea0bc82c656cddc39a87f2fd7313e722e",
"snapshot_id": "c7b4a227da49c51d14bb1f0b85f232d3391a91f3",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/enesusanmaz/azure-search-dotnet-samples/42d75a0ea0bc82c656cddc39a87f2fd7313e722e/create-first-app/README.md",
"visit_date": "2022-12-06T17:11:18.784081"
} | 3.796875 | stackv2 | ---
page_type: sample
languages:
- csharp
name: Create your first Azure Cognitive Search app
description: "Create a search page, and then enhance it with pagination controls, filters and facets, and typeahead queries. This example builds an ASP.NET Core MVC application using the Azure Cognitive Search .NET SDK."
products:
- azure
- azure-cognitive-search
urlFragment: create-first-app
---
# Create your first Azure Cognitive Search application

In this sample, start with a basic search page layout and then enhance it with paging controls, type-ahead (autocomplete), filtering and facet navigation, and results management.
This MVC sample is featured in [C# tutorial: Create your first app - Azure Cognitive Search](https://docs.microsoft.com/azure/search/tutorial-csharp-create-first-app). It's a collection of projects that demonstrate a user experience using fictitious hotels data. The first project creates a basic search page. Additional projects build on the first, adding results handling, and typeahead. The index is pre-built and hosted so that you can focus on the application itself.
## Contents
| File/folder | Description |
|-------------|-------------|
| `1-basic-search-page` | Initial project providing the page layout. |
| `2a-add-paging` | Adds a vertical scroll bar and page controls |
| `2b-add-infinite-scroll` | Demonstrates an infinite scrolling|
| `3-add-typeahead` | Adds autocomplete query |
| `4-add-facet-navigation` | Adds a facet navigation structure backed by filtering|
| `5-order-results` | Adds results sorting |
| `.gitignore` | Define what to ignore at commit time. |
| `CONTRIBUTING.md` | Guidelines for contributing to the sample. |
| `README.md` | This README file. |
| `LICENSE` | The license for the sample. |
## Prerequisites
- [Visual Studio 2019](https://visualstudio.microsoft.com/downloads/)
In contrast with other tutorials, this one uses a read-only hotels index on an existing demo search service maintained by Microsoft. No preliminary service or index setup is required.
## Setup
1. Clone or download this sample repository.
1. Extract contents if the download is a zip file. Make sure the files are read-write.
## Running create-first-app
Version 10 (v10) refers to the [Microsoft.Azure.Search](https://docs.microsoft.com/dotnet/api/overview/azure/search/client10) client library used in this code sample.
A newer version 11 client library ([Azure.Search.Documents](https://docs.microsoft.com/dotnet/api/overview/azure/search.documents-readme)) is now available. Over time, this code sample will be updated to include a v11 version, but currently only v10 is available at this time.
1. Open the \create-first-app\v10\1-basic-search-page folder
1. Open the FirstAzureSearchApp.sln project in Visual Studio
1. Compile and run the project
The solutions in this sample have template modifications. Methods in Startup.cs have been reordered, with app.UseCookiePolicy() relocated below app.UseMvc(...). This change addresses a known issue in .NET Core 2.x MVC apps where TempData is not persisted.
## Next steps
You can learn more about Azure Cognitive Search on the [official documentation site](https://docs.microsoft.com/azure/search). |
2024-11-18T18:03:44.956727+00:00 | 2021-09-07T15:07:12 | 8696e71810076cf307364f17e0985b13136e63cb | 4 | {
"blob_id": "8696e71810076cf307364f17e0985b13136e63cb",
"branch_name": "refs/heads/main",
"committer_date": "2021-09-07T15:07:12",
"content_id": "8b206d24a657a0bd97c3b33391a5245cfd3c260f",
"detected_licenses": [
"Apache-2.0"
],
"directory_id": "5ab1c89622f9cd98b5b8d9a588b09ab1740c1894",
"extension": "md",
"filename": "ADR-4.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": null,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 6788,
"license": "Apache-2.0",
"license_type": "permissive",
"path": "/adr/ADR-4.md",
"provenance": "stack-edu-markdown-0000.json.gz:45",
"repo_name": "rusriver/nats-architecture-and-design",
"revision_date": "2021-09-07T15:07:12",
"revision_id": "5ecaa9f428fa2a1ce42a86a27e01083221256f4b",
"snapshot_id": "b1bc1860812a0054821c1a8c4c3ece1b11c89585",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/rusriver/nats-architecture-and-design/5ecaa9f428fa2a1ce42a86a27e01083221256f4b/adr/ADR-4.md",
"visit_date": "2023-07-18T15:37:37.056426"
} | 4.0625 | stackv2 | # NATS Message Headers
|Metadata|Value|
|--------|-----|
|Date |2021-05-12|
|Author |@aricart, @scottf|
|Status |Partially Implemented|
|Tags |server, client|
## Context
This document describes NATS Headers from the perspective of clients. NATS
headers allow clients to specify additional meta-data in the form of headers.
The headers are effectively
[HTTP Headers](https://tools.ietf.org/html/rfc7230#section-3.2).
The salient points of the HTTP header specification are:
- Each header field consists of a case-insensitive field name followed by a
colon (`:`), optional leading whitespace, the field value, and optional
trailing whitespace.
- No spaces are allowed between the header field name and colon.
- Field value may be preceded or followed by optional whitespace.
- The specification may allow any number of strange things like comments/tokens
etc.
- The keys can repeat.
More specifically from [rfc822](https://www.ietf.org/rfc/rfc822.txt) Section
3.1.2:
> Once a field has been unfolded, it may be viewed as being composed of a
> field-name followed by a colon (":"), followed by a field-body, and terminated
> by a carriage-return/line-feed. The field-name must be composed of printable
> ASCII characters (i.e., characters that have values between 33. and 126.,
> decimal, except colon). The field-body may be composed of any ASCII
> characters, except CR or LF. (While CR and/or LF may be present in the actual
> text, they are removed by the action of unfolding the field.)
The only difference between a NATS header and HTTP is the first line. Instead of
an HTTP method followed by a resource, and the HTTP version (`GET / HTTP/1.1`),
NATS will provide a string identifying the header version (`NATS/X.x`),
currently 1.0, so it is rendered as `NATS/1.0␍␊`.
Please refer to the
[specification](https://tools.ietf.org/html/rfc7230#section-3.2) for information
on how to encode/decode HTTP headers.
### Enabling Message Headers
The server that is able to send and receive headers will specify so in it's
[`INFO`](https://docs.nats.io/nats-protocol/nats-protocol#info) protocol
message. The `headers` field if present, will have a boolean value. If the
client wishes to send headers, it has to enable it must add a `headers` field
with the `true` value in its
[`CONNECT` message](https://docs.nats.io/nats-protocol/nats-protocol#connect):
```
"lang": "node",
"version": "1.2.3",
"protocol": 1,
"headers": true,
...
```
### Publishing Messages With A Header
Messages that include a header have a `HPUB` protocol:
```
HPUB SUBJECT REPLY 23 30␍␊NATS/1.0␍␊Header: X␍␊␍␊PAYLOAD␍␊
HPUB SUBJECT REPLY 23 23␍␊NATS/1.0␍␊Header: X␍␊␍␊␍␊
HPUB SUBJECT REPLY 48 55␍␊NATS/1.0␍␊Header1: X␍␊Header1: Y␍␊Header2: Z␍␊␍␊PAYLOAD␍␊
HPUB SUBJECT REPLY 48 48␍␊NATS/1.0␍␊Header1: X␍␊Header1: Y␍␊Header2: Z␍␊␍␊␍␊
HPUB <SUBJ> [REPLY] <HDR_LEN> <TOT_LEN>
<HEADER><PAYLOAD>
```
#### NOTES:
- `HDR_LEN` includes the entire serialized header, from the start of the version
string (`NATS/1.0`) up to and including the ␍␊ before the payload
- `TOT_LEN` the payload length plus the HDR_LEN
### MSG with Headers
Clients will see `HMSG` protocol lines for `MSG`s that contain headers
```
HMSG SUBJECT 1 REPLY 23 30␍␊NATS/1.0␍␊Header: X␍␊␍␊PAYLOAD␍␊
HMSG SUBJECT 1 REPLY 23 23␍␊NATS/1.0␍␊Header: X␍␊␍␊␍␊
HMSG SUBJECT 1 REPLY 48 55␍␊NATS/1.0␍␊Header1: X␍␊Header1: Y␍␊Header2: Z␍␊␍␊PAYLOAD␍␊
HMSG SUBJECT 1 REPLY 48 48␍␊NATS/1.0␍␊Header1: X␍␊Header1: Y␍␊Header2: Z␍␊␍␊␍␊
HMSG <SUBJECT> <SID> [REPLY] <HDR_LEN> <TOT_LEN>
<PAYLOAD>
```
- `HDR_LEN` includes the entire serialized header, from the start of the version
string (`NATS/1.0`) up to and including the ␍␊ before the payload
- `TOT_LEN` the payload length plus the HDR_LEN
## Decision
Implemented and merged to master.
## Consequences
Use of headers is possible.
## Compatibility Across NATS Clients
The following is a list of features to insure compatibility across NATS clients
that support headers. Because the feature in Go client and nats-server leverage
the Go implementation as described above, the API used will determine how header
names are serialized.
### Case-sensitive Operations
In order to promote compatibility across clients, this section describes how
clients should behave. All operations are _case-sensitive_. Implementations
should provide an option(s) to enable clients to work in a case-insensitive or
format header names canonically.
#### Reading Values
`GET` and `VALUES` are case-sensitive operations.
- `GET` returns a `string` of the first value found matching the specified key
in a case-sensitive lookup or an empty string.
- `VALUES` returns a list of all values that case-sensitive match the specified
key or an empty/nil/null list.
#### Setting Values
- `APPEND` is a case-sensitive, and case-preserving operation. The header is set
exactly as specified by the user.
- `SET` and `DELETE` are case-sensitive:
- `DELETE` removes headers in case-sensitive operation
- `SET` can be considered the result of a `DELETE` followed by an `APPEND`.
This means only exact-match keys are deleted, and the specified value is
added under the specified key.
#### Case-insensitive Option
The operations `GET`, `VALUES`, `SET`, `DELETE`, `APPEND` in the presence of a
`case-insensitive` match requirement, will operate on equivalent matches.
This functionality is constrained as follows:
- `GET` returns the first matching header value in a case-insensitive match.
- `VALUES` returns the union of all headers that case-insensitive match. If the
exact key is not found, an empty/nil/null list is returned.
- `DELETE` removes the all headers that case-insensitive match the specified
key.
- `SET` is the combination of a case-insensitive `DELETE` followed by an
`APPEND`.
- `APPEND` will use the first matching key found and add values. If no key is
found, values are added to a key preserving the specified case.
Note that case-insensitive operations are only suggested, and not required to be
implemented by clients, specially if the implementation allows the user code to
easily iterate over keys and values.
### Multiple Header Values Serialization
When serializing, entries that have more than one value should be serialized one
per line. While the http Header standard, prefers values to be a comma separated
list, this introduces additional parsing requirements and ambiguity from client
code. HTTP itself doesn't implement this requirement on headers such as
`Set-Cookie`. Libraries, such as Go, do not interpret comma-separated values as
lists.
|
2024-11-18T18:03:44.998053+00:00 | 2018-11-19T21:44:35 | 3176c9702f1c3c6d697caa91aab6ee6ee6125d72 | 3 | {
"blob_id": "3176c9702f1c3c6d697caa91aab6ee6ee6125d72",
"branch_name": "refs/heads/master",
"committer_date": "2018-11-19T21:44:35",
"content_id": "07e31f028a34e3647894122f164a83759da8f3b4",
"detected_licenses": [
"BSD-3-Clause"
],
"directory_id": "ad51cf550b74222c2e6af5d2d1c4b3584dacac43",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": 158047225,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 2219,
"license": "BSD-3-Clause",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:46",
"repo_name": "mahery-a/ecards",
"revision_date": "2018-11-19T21:44:35",
"revision_id": "0b73bfef17cf91d18fd1175a911ae8e2eb528231",
"snapshot_id": "93a90dd75962afe5473e3bdf046c551b6a097167",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/mahery-a/ecards/0b73bfef17cf91d18fd1175a911ae8e2eb528231/README.md",
"visit_date": "2020-04-07T04:15:21.870838"
} | 2.953125 | stackv2 | # Ecards
Greeting ecards web application: Angular(front-end) + Spring(Rest API) + MySQL
## Front-end (Angular 6 + Bootstrap 4)
In the application.properties, you have to modify the following:
* properties related to JavaMail
* spring.mail.host = smtp.gmail.com
* spring.mail.username =<EMAIL_ADDRESS> * spring.mail.password = ******
* properties related to the MySQL database
* spring.datasource.url = jdbc:mysql://localhost:3306/db_ecards?autoReconnect=true&useUnicode=true&characterEncoding=UTF-8&allowMultiQueries=true&useSSL=false
* spring.datasource.username = *********
* spring.datasource.password = *********
_Keywords: Typescript/Javascript, Single page application, html5, Css3, Responsive design(Flexbox),
Component architecture, functional/reactive programming (rxjs), routing/navigation, testing(Jasmine), security(authentification(Oauth2)/authorization), validation, exception handling, caching, internationalization, Visual studio code_
## Back-end (REST Api Spring 5 + Hibernate + MySQL)
You'll need to install the following dependencies:
* https://www.npmjs.com/package/angular2_photoswipe
* https://www.npmjs.com/package/ngx-quill
* https://www.npmjs.com/package/ngx-pagination
* https://www.npmjs.com/package/ngx-cookie-service
Also in the environment.ts, you have to modify the following properties:
* apiUrl: 'http://localhost:8080', (this is the url to the Spring REST API)
* giphy_api_key: '************' (you can get the Giphy API key here: https://developers.giphy.com/docs/)
_Keywords: REST API, Java, SQL, xml/Json, Spring boot, tests unitaires(JUnit+Mockito), Tomcat,
maven, Eclipse (STS), documentation(Swagger), multithreading/concurrency, Programmation orientée objet, design patterns, Service oriented architecture, security(authentification(Oauth2)/authorization), validation, exception handling, caching, monitoring, internationalization, versioning, pagination/filtering/sorting, content-negotiation, JavaMail_
## Screenshots





|
2024-11-18T18:14:54.226221+00:00 | 2021-07-15T07:08:16 | f54a217ec9afef97831bc6c37960bd6ee3d73d24 | 3 | {
"blob_id": "f54a217ec9afef97831bc6c37960bd6ee3d73d24",
"branch_name": "refs/heads/main",
"committer_date": "2021-07-15T07:08:16",
"content_id": "3c9dd7f00e959492b55ee5edadde2ba3bd449d1a",
"detected_licenses": [
"MIT",
"Apache-2.0"
],
"directory_id": "8c6be4bdf6a305b244b2209b2d4cdb344aafc9ec",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": "2020-07-13T09:55:42",
"gha_event_created_at": "2020-07-13T09:55:43",
"gha_language": null,
"gha_license_id": "Apache-2.0",
"github_id": 279264071,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 623,
"license": "MIT,Apache-2.0",
"license_type": "permissive",
"path": "/packages/grafana-toolkit/node_modules/@types/connect-history-api-fallback/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:47",
"repo_name": "ppp09812345/grafana",
"revision_date": "2021-07-15T07:08:16",
"revision_id": "c93d310ba9c1158cf38568e45a6a7e44c508c4f2",
"snapshot_id": "520c55e936f9d3d9208b7166261e90b82a9eb0d3",
"src_encoding": "UTF-8",
"star_events_count": 1,
"url": "https://raw.githubusercontent.com/ppp09812345/grafana/c93d310ba9c1158cf38568e45a6a7e44c508c4f2/packages/grafana-toolkit/node_modules/@types/connect-history-api-fallback/README.md",
"visit_date": "2023-06-16T16:49:16.813823"
} | 2.53125 | stackv2 | # Installation
> `npm install --save @types/connect-history-api-fallback`
# Summary
This package contains type definitions for connect-history-api-fallback (https://github.com/bripkens/connect-history-api-fallback#readme).
# Details
Files were exported from https://github.com/DefinitelyTyped/DefinitelyTyped/tree/master/types/connect-history-api-fallback
Additional Details
* Last updated: Mon, 19 Aug 2019 00:51:09 GMT
* Dependencies: @types/express-serve-static-core, @types/node
* Global values: none
# Credits
These definitions were written by Douglas Duteil <https://github.com/douglasduteil>.
|
2024-11-18T18:14:54.265952+00:00 | 2021-02-24T22:50:13 | a374da4b6fc28c27e0cf1c90780b12fb622b57ba | 2 | {
"blob_id": "a374da4b6fc28c27e0cf1c90780b12fb622b57ba",
"branch_name": "refs/heads/master",
"committer_date": "2021-02-24T22:50:13",
"content_id": "df3ef98ae55ec4c97bec06d7f07de0212a373313",
"detected_licenses": [
"MIT"
],
"directory_id": "9262e021325882c80a1f73234782ef9297ca6e68",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": "2019-10-26T20:26:02",
"gha_event_created_at": "2020-05-28T15:27:44",
"gha_language": "Java",
"gha_license_id": null,
"github_id": 217765300,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 242,
"license": "MIT",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:48",
"repo_name": "ThomasJRichards17/WordsearchSolver",
"revision_date": "2021-02-24T22:50:13",
"revision_id": "6558f52486ce6d90112af49b3fda2cd9a1bb1669",
"snapshot_id": "8d8827197c781257dea926366c6fd84e39469cf0",
"src_encoding": "UTF-8",
"star_events_count": 1,
"url": "https://raw.githubusercontent.com/ThomasJRichards17/WordsearchSolver/6558f52486ce6d90112af49b3fda2cd9a1bb1669/README.md",
"visit_date": "2023-03-12T22:16:52.130969"
} | 2.03125 | stackv2 | # WordsearchSolver
- Android app to solve wordsearches from images as well as manual text input
- Uses Firebase Machine Learning Vision to detect words from images
- Allows you to save solved wordsearches to their device and load them later
|
2024-11-18T18:20:58.400264+00:00 | 2015-10-25T23:32:25 | f88846a973635b605f0142eec9ac4602192f6081 | 4 | {
"blob_id": "f88846a973635b605f0142eec9ac4602192f6081",
"branch_name": "refs/heads/master",
"committer_date": "2015-10-25T23:32:25",
"content_id": "07d2cb3c1a5ddfc9f2dfa8b7c7a857ad7df39998",
"detected_licenses": [
"Apache-2.0"
],
"directory_id": "02343f22c2510e13fb537051daa123cff82c3b5c",
"extension": "md",
"filename": "CHANGELOG.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": null,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 2297,
"license": "Apache-2.0",
"license_type": "permissive",
"path": "/CHANGELOG.md",
"provenance": "stack-edu-markdown-0000.json.gz:49",
"repo_name": "eriknyk/react-dom-stream",
"revision_date": "2015-10-25T23:32:25",
"revision_id": "ffb811086f91e1364a78e103d525194e7b5de741",
"snapshot_id": "4a33904dd819e74c3f13724a2d98248b6994f825",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/eriknyk/react-dom-stream/ffb811086f91e1364a78e103d525194e7b5de741/CHANGELOG.md",
"visit_date": "2021-01-21T07:39:55.760987"
} | 3.5 | stackv2 | ## v0.2.0
This version's main achievement is changing the API to be more stream-friendly. The 0.1.x API is still supported, but it is deprecated and will cause a console error. In version 0.3.0, I will remove support for the 0.1.x API.
### Converting code from the v0.1.x API to the v0.2.x API
The first difference between v0.1.x's API and v0.2.x's API is how they handle the stream. v0.1.x accepted a Writable stream as an argument to `renderToString` and `renderToStaticMarkup`, but v0.2.x instead returns a Readable stream.
The second difference is that the hash Promise returned from `renderToString` is now a property called `hash` on the stream that is the return value.
The third difference is that there is no longer an `options` argument for either server-side render method, and the methods no longer buffer their output. Buffering is still vital to getting good performance, but you can (and should!) use projects like `compression` or `buffered-stream`. For more information, see the buffering section of the readme.
If your `renderToString` code looks like this in v0.1.x:
```javascript
var ReactDOMStream = require("react-dom-stream/server");
app.get('/', function (req, res) {
ReactDOMStream.renderToString(<Foo prop={value}/>, res)
.then(function(hash) {
// TODO: write the hash out to the page in a script tag
res.end();
});
});
```
Then it should look like this in v0.2.x:
```javascript
var ReactDOMStream = require("react-dom-stream/server");
app.get('/', function (req, res) {
var stream = ReactDOMStream.renderToString(<Foo prop={value}/>);
stream.pipe(res, {end: false});
stream.hash.then(function(hash) {
// TODO: write the hash out to the page in a script tag
res.end();
});
});
```
Since `renderToStaticMarkup returns a stream without a hash parameter, its code is much simpler. The following v0.1.x code:
```javascript
var ReactDOMStream = require("react-dom-stream/server");
app.get('/', function (req, res) {
ReactDOMStream.renderToStaticMarkup(<Foo prop={value}/>, res)
.then(function() {
res.end();
});
});
```
looks like this in v0.2.x:
```javascript
var ReactDOMStream = require("react-dom-stream/server");
app.get('/', function (req, res) {
ReactDOMStream.renderToStaticMarkup(<Foo prop={value}/>).pipe(res);
});
```
|
2024-11-18T18:26:03.155455+00:00 | 2019-03-30T00:37:37 | 32702a4b514dd7e2bbbbb2c862a1befebe81189b | 3 | {
"blob_id": "32702a4b514dd7e2bbbbb2c862a1befebe81189b",
"branch_name": "refs/heads/master",
"committer_date": "2019-03-30T00:37:37",
"content_id": "d6f295058020c6e8f81b5f4768ceac1bba7d77e8",
"detected_licenses": [
"MIT"
],
"directory_id": "2341b6f541f906ddce28835bec699eb30281e128",
"extension": "markdown",
"filename": "2014-09-05-conversational-developer-interview-questions.markdown",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": 35223724,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 4747,
"license": "MIT",
"license_type": "permissive",
"path": "/_posts/2014-09-05-conversational-developer-interview-questions.markdown",
"provenance": "stack-edu-markdown-0000.json.gz:50",
"repo_name": "dummied/dummied.github.io",
"revision_date": "2019-03-30T00:37:37",
"revision_id": "449ce2f68f00fca1f487a7f9fd79c5ed3c1a5e0f",
"snapshot_id": "0ffbac0d41403847128e71a9cd0a4a1ca8dbd679",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/dummied/dummied.github.io/449ce2f68f00fca1f487a7f9fd79c5ed3c1a5e0f/_posts/2014-09-05-conversational-developer-interview-questions.markdown",
"visit_date": "2021-01-01T05:34:40.386365"
} | 3.15625 | stackv2 | ---
layout: post
title: Conversational developer interview questions
date: '2014-09-05 16:02:41'
---
I've [already recommended not doing interviews at all](http://blog.chrisvannoy.com/book-preview-the-un-interview/), but rather conversations.
But you're still going to need to ask questions. You still need some way to ensure you backstop all of your research.
The other thing I wouldn't recommend at this point are your traditional interview questions that are supposed to gain some insight into the candidate's psyche. They're not going to fit in with our un-interview vibe and aren't useful anyway.
As far as what type of question I'm referring to, I mean the "If you were an animal, what kind of animal would you be?"
I worked with a digital hiring manager a while back that would ask every candidate, "If you could have one super-power, what would it be?"
He had meanings and talismans associated with all manner of answers.
* **Omniscience:** Definite red flag. Too much desire for power. Will be hard to work with.
* **Invisibility:** Similar red flag. Wants to shrink away. Downside: this is a popular developer answer.
Like we’ve already talked about, you only need to care about two things: Can they do the job and can you work with them? That’s it.
So, what sorts of questions should you ask?
Below are some examples, in rough conversational order:
* **Where are you working these days? Do you like it?**
*What you’re looking for:* An in. This question has less to do with how they’ll fit in, but much more to do with how you can get them in. You’re looking for contrast points with what you have to offer. What are their pain points and what can you sell to them. Alternatively, if they’re not complaining at all, you might have to do a lot more work later. That said: Even the happiest developer will complain about *something* if you give them a window. Use that later.
* **What do they have you working on? How are you building that?**
*What you’re looking for:* A bit of the above — job dissatisfaction, but also a segue into later questions. Also look for connections, thematically or otherwise, with the work you have in mind for them. They don’t have to be connected, but if they are, that can be a handy indicator that they can do the job you’re hiring for — because they’re already kinda doing it.
* **What’s your favorite language? Why?**
*What you’re looking for:* You don’t necessarily need your programming language to be the favorite. Look more for the energy behind this question. They should be excited to talk about languages and should have enough knowledge of their favorite and others to give good reasons for liking it. The actual content of the answer doesn’t matter nearly as much as those factors: are they excited when they talk about languages, and do they seem somewhat familiar with more than one?
* **Anything on the side?**
*What you’re looking for:* Ideally something. Side projects are another important indicator. The great developers are always exploring, fiddling and screwing around with experiments outside their day job. What that side project(s) consists of can also give you a glimpse at what they truly care about.
* **Any experience with X?**
*What you’re looking for:* X in this case is some specific requirement or nice-to-have you have in mind. If they say no, be prepared to explain what it is you’re talking about to see if they at least get it conceptually.
* **What do you do for fun?**
*What you’re looking for:* This one can go two ways. If the answer is “more programming," that’s a good sign (see the Anything on the side question above). But it’s not a negative if that’s not the answer. After all, you want to hire interesting people who like to explore. You want to get a sense of that. As an aside: It’s been my experience that home brewing beer is a popular answer to this, at least among the best developers I've known or hired.
* **What’s the nastiest problem you’ve run across?**
*What you’re looking for:* Almost every developer has a horror story or two. You want to hear how they thought through the problem and how it turned out.
The important thing to keep in mind with any of these is that they aren’t interview questions, and again, this isn’t an interview. It’s a conversation, preferably over food or beverage. Talk about yourself, talk about your company, let these questions lead off into tangents and rabbit holes — just like a regular conversation, because that’s what it is.
Always keep in mind the three key questions you're wanting answered for yourself:
"Can this person do the job?"
"Can I work them?"
"Would they like working here?" |
2024-11-18T18:26:03.195398+00:00 | 2019-09-16T21:19:36 | 147c1afed65dffda2cf8e4061c957b94ffc5fef8 | 3 | {
"blob_id": "147c1afed65dffda2cf8e4061c957b94ffc5fef8",
"branch_name": "refs/heads/master",
"committer_date": "2019-09-16T21:19:36",
"content_id": "a371cb1ae70147eb8cf5ff5d2d094dd12350e36a",
"detected_licenses": [
"MIT"
],
"directory_id": "e8d7e16d4882e426239e4f7353e55cfeeec5880d",
"extension": "md",
"filename": "_index.md",
"fork_events_count": 0,
"gha_created_at": "2019-06-27T14:16:03",
"gha_event_created_at": "2019-06-27T14:16:04",
"gha_language": null,
"gha_license_id": "MIT",
"github_id": 194108371,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 1887,
"license": "MIT",
"license_type": "permissive",
"path": "/docs/content/docs/storefront/_index.md",
"provenance": "stack-edu-markdown-0000.json.gz:51",
"repo_name": "lsimone/vendure",
"revision_date": "2019-09-16T21:19:36",
"revision_id": "578624bb7e68d41db10d9c266e37b1a6673b3049",
"snapshot_id": "c8bc5bda1f7f4a2e71755b41288cb356cae99640",
"src_encoding": "UTF-8",
"star_events_count": 1,
"url": "https://raw.githubusercontent.com/lsimone/vendure/578624bb7e68d41db10d9c266e37b1a6673b3049/docs/content/docs/storefront/_index.md",
"visit_date": "2020-06-11T22:36:07.795503"
} | 3.3125 | stackv2 | ---
title: "Storefront"
weight: 1
---
# Storefront
The storefront is the application which customers use to interact with your store.
As a headless server, Vendure provides a GraphQL API and admin UI app, but no storefront. The key advantage of the headless model is that the storefront (or indeed, any number of client applications) can be developed completely independently of the server. This flexibility comes at the cost of having to build and maintain your own storefront.
However, we'd like to lower the barrier to getting started in the regard, so there are plans for integrations with existing e-commerce storefront solutions as well as our own vendure-storefront project.
## Vendure Storefront
{{< figure src="./vendure-storefront-screenshot-01.jpg" >}}
This is our own progressive web application (PWA) storefront. It is currently still in development, but when complete it will offer a ready-made, customizable storefront solution tailored specifically to work well with the Vendure server.
A live demo can be found here: [demo.vendure.io/storefront/](https://demo.vendure.io/storefront/)
Keep up with development here: [github.com/vendure-ecommerce/storefront](https://github.com/vendure-ecommerce/storefront)
## DEITY Falcon
[DEITY Falcon](https://falcon.deity.io/docs/getting-started/intro) is a React-based PWA storefront solution. It uses a modular architecture which allows it to connect to any e-commerce backend. We are developing the [Vendure Falcon API](https://www.npmjs.com/package/@vendure/falcon-vendure-api) which allows Falcon to be used with Vendure.
Here's a video showing how to quickly get started with Vendure + DEITY Falcon:
{{< vimeo 322812102 >}}
## Gatsby
We are also working on a [Gatsby](https://www.gatsbyjs.org/)-based storefront app: [vendure-ecommerce/gatsby-storefront](https://github.com/vendure-ecommerce/gatsby-storefront)
|
2024-11-18T18:26:03.287149+00:00 | 2021-04-09T15:30:02 | e07b2c7176c2dbd3aca0caf73dc85210abc597ec | 3 | {
"blob_id": "e07b2c7176c2dbd3aca0caf73dc85210abc597ec",
"branch_name": "refs/heads/master",
"committer_date": "2021-04-09T15:30:02",
"content_id": "7d92947020082b5a35b138229503a4944031fdcd",
"detected_licenses": [
"Apache-2.0"
],
"directory_id": "81c3432d1aa5c112b74dc712172863088d2edc33",
"extension": "md",
"filename": "run.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": 332756842,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 520,
"license": "Apache-2.0",
"license_type": "permissive",
"path": "/doc/run.md",
"provenance": "stack-edu-markdown-0000.json.gz:52",
"repo_name": "space-sh/sshd",
"revision_date": "2021-04-09T15:30:02",
"revision_id": "38fca8630d8e66ee88998fa6b76d353ddaf7cf94",
"snapshot_id": "ff23958bfd9c988bed4fb74978dc1c3049d8c5e4",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/space-sh/sshd/38fca8630d8e66ee88998fa6b76d353ddaf7cf94/doc/run.md",
"visit_date": "2023-03-29T20:04:58.659737"
} | 3.359375 | stackv2 | ---
modulename: SSHD
title: /run/
giturl: gitlab.com/space-sh/sshd
editurl: /edit/master/doc/run.md
weight: 200
---
# SSHD module: Run
Run a local _SSHD_ server.
## Example
First generate the server host key:
```sh
space -m sshd /config/
```
Default _SSHD_ server start:
```sh
space -m sshd /run/
```
_SSHD_ server start on port 22 and customized file paths:
```sh
space -m sshd /run/ -- "/home/janitor/ssh_host_rsa_key" "22" "/home/janitor/authorized_keys"
```
Exit status code is expected to be 0 on success.
|
2024-11-18T18:26:03.322475+00:00 | 2017-07-16T14:56:30 | 95f69902980c9e4f348a3996fab3b9d52a68f3e7 | 3 | {
"blob_id": "95f69902980c9e4f348a3996fab3b9d52a68f3e7",
"branch_name": "refs/heads/1.3.1",
"committer_date": "2017-07-16T14:56:46",
"content_id": "1afce29f534d54eb1b2fc2b8abb19eab7a8ffffd",
"detected_licenses": [
"MIT"
],
"directory_id": "6529e9a449909b7724e9a00bb56ea2c6c130e71d",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": "2017-09-30T06:17:01",
"gha_event_created_at": "2017-09-30T06:17:01",
"gha_language": null,
"gha_license_id": null,
"github_id": 105346838,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 6229,
"license": "MIT",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:53",
"repo_name": "TtbqSora/ray-mmd",
"revision_date": "2017-07-16T14:56:30",
"revision_id": "f55614750c377cfe751b4f393a9bb8f047ac05a3",
"snapshot_id": "5a8ea75122c6aa8bd073b23f01c3c8345ed6bc6d",
"src_encoding": "UTF-8",
"star_events_count": 3,
"url": "https://raw.githubusercontent.com/TtbqSora/ray-mmd/f55614750c377cfe751b4f393a9bb8f047ac05a3/README.md",
"visit_date": "2021-07-09T07:14:58.584065"
} | 3.046875 | stackv2 | Ray-MMD
========
### Physically-Based Rendering ###
The aim of the project is to create a physically-based rendering at MMD.
Screenshots :
------------
[](https://github.com/ray-cast/images/raw/master/screen1.jpg)
[](https://github.com/ray-cast/images/raw/master/screen2.png)
[](https://github.com/ray-cast/images/raw/master/hou.jpg)
[](https://github.com/ray-cast/images/raw/master/lights.png)
Download :
------------
* [Ray-MMD 1.3.1 - Source code (zip)](https://github.com/ray-cast/ray-mmd/archive/1.3.1.zip) (updated: 27. Apr 2017)
* [Ray-MMD 1.3.1 - Source code (tar.gz)](https://github.com/ray-cast/ray-mmd/archive/1.3.1v2.tar.gz) (updated: 27. Apr 2017)
- Latest changes - Major allocator changes:
- Added a LightBloom Extension
- Optimized for cloth lighting
- Fixed bug : when set NORMAL_MAP_FROM to 9
- Fixed bug : when add Sub Normal Map to editor.fx
Requirement :
------------
* MikuMikuDance - 926ver and above (Without Anti-Aliasing)
* MikuMikuEffect - 037ver and above
* Direct3D 9 With Shader Model 3.0 (ps_3_0)
Resources
------------
- HDRi
- sIBL Archive - Hdrlabs.com \[[link](http://www.hdrlabs.com/sibl/archive.html)\].
- ++skies; - **[aokcub](https://twitter.com/aokcub_cg)** \[[link](https://aokcub.net/cg/incskies/)\].
- USC Institute \[[link](http://gl.ict.usc.edu/Data/HighResProbes)\].
- Text editor
- Notepad++ \[[link](https://notepad-plus-plus.org)\].
- Visual studio code \[[link](http://code.visualstudio.com/Download)\].
- Materials
- Hair for Apperience Models - by VanillaBear3600 \[[link](http://vanillabear3600.deviantart.com/art/RayCast-Hair-Shader-For-Apperience-Models-664061177)\].
Tutroial:
------------
* Chinese Pages \[[link](https://github.com/ray-cast/ray-mmd/wiki/0.0-%E6%95%99%E7%A8%8B)\].
Features :
------------
* Physically-Based Material
* Multiple Light Source
* IES Light Profiles
* Image Based Lighting
* Scene Space Reflection
* Screen Space Ambient Occlusition
* Screen Space Subsurface Scattering
* Color Balance PostProcess
* HDR PostProcess
* Bloom PostProcess
* FXAA PostProcess
* SMAA PostProcess
Contact
------------
* Reach me via Twitter: [@Rui](https://twitter.com/Rui_cg).
[License (MIT)](https://raw.githubusercontent.com/ray-cast/ray-mmd/developing/LICENSE.txt)
-------------------------------------------------------------------------------
Copyright (C) 2016-2017 Ray-MMD Developers. All rights reserved.
https://github.com/ray-cast/ray-mmd
Permission is hereby granted, free of charge, to any person obtaining a
copy of this software and associated documentation files (the "Software"),
to deal in the Software without restriction, including without limitation
the rights to use, copy, modify, merge, publish, distribute, sublicense,
and/or sell copies of the Software, and to permit persons to whom the
Software is furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be included
in all copies or substantial portions of the Software.
IN NO EVENT SHALL
BRIAN PAUL BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN
AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
Digging Deeper
--------
* Moving to the Next Generation - The Rendering Technology of Ryse \[[link](http://www.crytek.com/download/2014_03_25_CRYENGINE_GDC_Schultz.pdf)\].
* ACES Filmic Tone Mapping Curve \[[link](https://knarkowicz.wordpress.com/2016/08/31/hdr-display-first-steps/)\].
* Compact Normal Storage for small G-Buffers \[[link](http://aras-p.info/texts/CompactNormalStorage.html)\].
* Convert Blinn-Phong to Beckmann distribution \[[link](http://simonstechblog.blogspot.de/2011/12/microfacet-brdf.html)\].
* Spherical Gaussian approximation for Blinn-Phong, Phong and Fresnel \[[link](https://seblagarde.wordpress.com/2012/06/03/spherical-gaussien-approximation-for-blinn-phong-phong-and-fresnel/)\].
* Physically Based Area Lights \[[link](http://www.frostbite.com/wp-content/uploads/2014/11/course_notes_moving_frostbite_to_pbr.pdf)\].
* Physics and Math of Shading \[[link](http://blog.selfshadow.com/publications/s2015-shading-course/hoffman/s2015_pbs_physics_math_slides.pdf)\].
* Compact YCoCg Frame Buffer for small IBL-Buffer \[[link](http://jcgt.org/published/0001/01/02/)\].
* RGBM color encoding \[[link](http://graphicrants.blogspot.com/2009/04/rgbm-color-encoding.html)\].
* Horizon Occlusion for IBL \[[link](http://marmosetco.tumblr.com/post/81245981087)\].
* Screen space glossy reflections \[[link](http://roar11.com/2015/07/screen-space-glossy-reflections/)\].
* Parallax Occlusion Map \[[link](http://sunandblackcat.com/tipFullView.php?topicid=28)\].
* Convert Temperature to RGB \[[link](https://github.com/davidf2281/ColorTempToRGB)\].
* Texture repetition \[[link](http://www.iquilezles.org/www/articles/texturerepetition/texturerepetition.htm)\].
* Pre-Integrated Skin Shading \[[link](http://simonstechblog.blogspot.com/2015/02/pre-integrated-skin-shading.html)\]
* Normal Blending in Detail \[[link](http://blog.selfshadow.com/publications/blending-in-detail/)\]
* An Approximation to the Chapman Grazing-Incidence Function for Atmospheric Scattering \[[link](http://www.gameenginegems.net/gemsdb/article.php?id=1133)\]
* Bump map to normal \[[link](https://docs.unrealengine.com/latest/attachments/Engine/Rendering/LightingAndShadows/BumpMappingWithoutTangentSpace/mm_sfgrad_bump.pdf)\]
* Special-Case Materials Wetness \[[link](http://advances.realtimerendering.com/other/2016/naughty_dog/NaughtyDog_TechArt_Final.pdf)\] |
2024-11-18T18:26:03.367486+00:00 | 2019-11-07T16:52:12 | e7bc03e3e3ae6ea9eda311f3f820939ee0857c58 | 4 | {
"blob_id": "e7bc03e3e3ae6ea9eda311f3f820939ee0857c58",
"branch_name": "refs/heads/master",
"committer_date": "2019-11-07T16:52:12",
"content_id": "c46028bae04ed13334ef4162ffd0d4aba2787e17",
"detected_licenses": [
"Apache-2.0"
],
"directory_id": "dd8e37990a9743f528c4bc0ee31bfac102a41da5",
"extension": "md",
"filename": "README.md",
"fork_events_count": 6,
"gha_created_at": "2018-03-07T08:47:05",
"gha_event_created_at": "2022-10-06T02:34:16",
"gha_language": "Go",
"gha_license_id": "Apache-2.0",
"github_id": 124206479,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 15157,
"license": "Apache-2.0",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:54",
"repo_name": "coinbase/step-asg-deployer",
"revision_date": "2019-11-07T16:52:12",
"revision_id": "e131cc675339259d6a5be8366fdb172e4c6cd06b",
"snapshot_id": "2d1a0d88fe7074234d05c5a3f24a10ad3262f2e2",
"src_encoding": "UTF-8",
"star_events_count": 10,
"url": "https://raw.githubusercontent.com/coinbase/step-asg-deployer/e131cc675339259d6a5be8366fdb172e4c6cd06b/README.md",
"visit_date": "2022-10-22T23:48:14.206430"
} | 3.96875 | stackv2 | # Step Auto Scaling Group (ASG) Deployer a.k.a. Asgard
<img src="./assets/sad-logo.png" align="right" alt="Asgard" />
Deploy your [12-factor-applications](https://12factor.net/) to AWS easily and securely with the [Step](https://github.com/coinbase/step) [Auto-Scaling Group](https://docs.aws.amazon.com/autoscaling/ec2/userguide/what-is-amazon-ec2-auto-scaling.html) (ASG) Deployer (Asgard).
Asgard's goals/requirements/features are:
1. **Ephemeral Blue/Green**: create new instances, wait for them to become healthy, delete old instances.
1. **Declarative**: describe what a successful release looks like, not how to deploy it.
1. **Scalable**: can scale both vertically (larger instances) and horizontally (more instances).
1. **Secure**: resources are verified to ensure that they cannot be used accidentally or maliciously.
1. **Gracefully Fail**: handle failures to recover and roll back with no/minimal impact to users.
1. **Configuration Parity**: minimize divergence between production, staging and development environments by keeping releases as similar as possible.
1. **Cattle not Pets**: treat compute instances as disposable and ephemeral.
1. **No Deployer Configuration**: no configuration and minimal setup needed to get Asgard up and running.
1. **Multi Account**: one deployer for all AWS accounts.
### Getting Started
Asgard is made of an [AWS Lambda Function](https://docs.aws.amazon.com/lambda/latest/dg/welcome.html) (with a role) and [AWS Step Function](https://docs.aws.amazon.com/step-functions/latest/dg/welcome.html). You can bootstrap these into AWS with:
```bash
git pull # pull down new code
./scripts/bootstrap
```
#### Testing Asgard with deploy-test
Asgard includes a test project `deploy-test` that has one service `web` which is a nginx server to be mounted behind a [Elastic Load Balancer](https://aws.amazon.com/elasticloadbalancing/) (ELB) and [Load Balancer](https://docs.aws.amazon.com/elasticloadbalancing/latest/application/introduction.html) target group. The service instances have a [security group](https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/using-network-security.html) and [instance profile](https://docs.aws.amazon.com/IAM/latest/UserGuide/id_roles_use_switch-role-ec2_instance-profiles.html).
To create the AWS resources for `deploy-test`:
```bash
./scripts/geo apply resources/deploy-test-resources.rb
```
*Note: you will also have to tag the latest Ubuntu release*
A `deploy-test` release file `deployer-test-release.json` looks like:
```yaml
{
"project_name": "coinbase/deploy-test",
"config_name": "development",
"subnets": ["test_private_subnet_a", "test_private_subnet_b"],
"ami": "ubuntu",
"user_data": "{{USER_DATA_FILE}}",
"services": {
"web": {
"instance_type": "t2.nano",
"security_groups": ["ec2::coinbase/deploy-test::development"],
"elbs": ["coinbase-deploy-test-web-elb"],
"profile": "coinbase-deploy-test",
"target_groups": ["coinbase-deploy-test-web-tg"]
}
}
}
```
The user data for the release is stored in the file `deployer-test-release.json.userdata`:
```yaml
#cloud-config
repo_update: true
repo_upgrade: all
packages:
- docker.io
runcmd:
- docker run -d --restart always --name test_server -p 8000:80 nginx
```
To build a release for `deploy-test` and send it to Asgard we use the `step-asg-deployer` executable:
```bash
step-asg-deployer deploy deploy-test-release.json
```
<img src="./assets/sad-deploy.gif" alt="Asgard deploy" />
Asgard then:
1. validates the sent release and any referenced resources.
1. creates a new auto-scaling group for `web` which is configured to start an nginx server.
1. waits for the EC2 instances in the `web` ASG to become healthy behind the ELB and target group. Healthy means that the health checks for both ELB and target group pass.
1. Once healthy the old ASG and its instances are terminated.
### Asgard Release
An Asgard release is a request to deploy a **Project-Configuration** where:
* A **Project** is a code-base typically named with `org/name`.
* A **Configuration** is the environment the project is being deployed into, e.g. `development`, `production`.
Each release can define 1-to-many **Services**; each service is a logical group of servers, e.g. `web` or `worker`, that maps to a single auto-scaling group (ASG).
When Asgard is sent a release, it moves it through a state machine:
<img src="./assets/sad-success.png" alt="sad state diagram"/>
1. **Validate**: validate the release is correct.
1. **Lock**: grabs a lock on project-configuration.
1. **ValidateResources**: validate resources w.r.t. the project, configuration and service using them.
1. **Deploy**: creates an ASG and other resource for each service.
1. **CheckHealthy**: check to see if the new instances created are healthy w.r.t. their ASGs ELBs and target groups. If instances are seen to be terminating immediately halt release.
1. **CleanUpSuccess**: if the release was a success, then delete the old ASGs.
1. **CleanUpFailure**: if the release failed, delete the new ASGs.
1. **ReleaseLockFailure**: try to release the lock and fail.
At each of these states it is possible to fail and then move towards a failure state. The typical failures are:
* **BadReleaseError**: The release sent was invalid because either its structure was incorrect, its values were invalid, or its resources were invalid.
* **LockExistsError**: Could not grab the lock because either another deploy for the project-configuration is currently going out, or a previous deploy left a lock in place.
* **DeployError**: Unable to create a new ASG or resource.
* **HaltError**: Halt was detected or instances were found terminating.
* **TimeoutError**: The deploy took too long and failed.
The end states are:
1. **Success**: the release went went as planned.
2. **FailureClean**: release was unsuccessful, but cleanup was successful, so AWS was left in good state.
3. **FailureDirty**: release was unsuccessful, but cleanup failed so AWS was left in a bad state. This should never happen and should alert if this happens, and file a bug.
4. It is possible to not end in one of these states if the state machine is incorrect. **This is very bad**, alert if this happens and file a bug.
#### Resources
A release uses resources that must exist and be configured correctly to be used for the project-configuration-service being deployed.
A release **must** have:
1. an **AMI** defined with the `ami` key that can be either a `Name` tag or AMI ID e.g. `ami-1234567`
2. **Subnets** defined with `subnets` key that is a list of either `Name` tags or Subnet IDs e.g. `subnet-1234567`
Both the above resources **MUST** have a tag `DeployWith` that equals `step-asg-deployer`.
Services **can** have:
1. **Security Groups** defined with `security_groups` key is a list of security groups `Name` tags
2. **Elastic Load Balancers** defined with `elbs` key is a list of ELB names
3. **Application Load Balancer Target Groups** defined with `target_groups` is a list of target group's `Name` tags
All the above resources **MUST** be tagged with the `ProjectName`, `ConfigName` and `ServiceName` of the release to ensure that resources are assigned correctly.
Services can also have an **Instance Profile** defined by the `profile` key that is and instance profile `Name` tag. The roles path **MUST** be equal to `/<project_name>/<config_name>/<service_name>/`.
#### Scale
Asgard makes it easy to scale both vertically and horizontally. To scale `deploy-test` we add to the release:
```yaml
{ ...
"services": {
"web": { ...
"instance_type": "c4.xlarge",
"ebs_volume_size": 20,
"ebs_volume_type": "gp2",
"ebs_device_name": "/dev/sda1",
"autoscaling": {
"min_size": 3,
"max_size": 5,
"spread": 0.2,
"max_terms": 1,
"policies": [
{
"type": "cpu_scale_up",
"threshold" : 25,
"scaling_adjustment": 2
},
{
"type": "cpu_scale_down",
"threshold" : 15,
"scaling_adjustment": -1
}
]
}
}
}
}
```
* `instance_type` is the [EC2 instance type](https://www.ec2instances.info/) for the service
* `ebs_volume_size`, `ebs_volume_type`, `ebs_device_name` define the attached [EBS volume](https://aws.amazon.com/ebs/) in GB.
The `autoscaling` key defines the horizontal scaling of a service:
* all calculations are bounded by `min_size` and `max_size`.
* the `desired_capacity` is equal to the `min_size` or capacity of the previously launched service
* the actual number of instances launched is the `desired_capacity * (1 + spread)`
* to be deemed the healthy the service must have `desired_capacity * (1 - spread)`
* if the number of terminating is greater than or equal to `max_terms` (default `0`), the release is immediately halts.
* `policies` are defined above to increase the `desired_capacity` by 2 instances if the CPU goes above 25% and reduce by 1 instance if it drops below 15%.
*Both `spread` and `max_terms` are useful when launching many instances because as scale increases the number of cloud errors increase.*
#### User Data
**Do not put sensitive data into user data**. User data is not treated by Asgard as secure information, it is difficult to secure with IAM, and it is very [limited in size](https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/ec2-instance-metadata.html#instancedata-add-user-data). We recommend using [Vault](https://www.vaultproject.io/), [AWS Parameter store](https://docs.aws.amazon.com/systems-manager/latest/userguide/systems-manager-paramstore.html), or [KMS encrypted S3](https://docs.aws.amazon.com/kms/latest/developerguide/services-s3.html) authenticated by a service's instance profile.
The `user_data` in the release is the plain text [instance metadata](https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/ec2-instance-metadata.html) sent to initialize each instance. Asgard will replace some strings with information about the release, project, config and service, e.g.:
```yaml
...
write_files:
- path: /
content: |
{{RELEASE_ID}}
{{PROJECT_NAME}}
{{CONFIG_NAME}}
{{SERVICE_NAME}}
```
Asgard will replace `{{PROJECT_NAME}}` with the name of the project and `{{SERVICE_NAME}}` with the name of the service. This can be useful for getting service specific configuration and logging.
If `user_data` is equal to `{{USER_DATA_FILE}}` and deployed with `step-asg-deployer` the value will be replaced with the contents of the `<release_file>.userdata`, e.g. `deployer-test-release.json.userdata`.
#### Timeout
A release can have a `timeout` which is how long in seconds a release will wait for its services to become healthy. By default the timeout is 10 minutes, the max value would be around a year (*31556926 seconds*) since that is how long a step function can run.
#### Lifecycle
AWS provides [Auto Scaling Group Lifecycle Hooks](https://docs.aws.amazon.com/autoscaling/ec2/userguide/lifecycle-hooks.html) to detect and react to auto-scaling events. You can add the lifecycle hooks to the ASGs with:
```yaml
{ ...
"lifecycle": {
"termhook" : {
"transition": "autoscaling:EC2_INSTANCE_TERMINATING",
"role": "asg_lifecycle_hooks",
"sns": "asg_lifecycle_hooks",
"heartbeat_timeout": 300
}
}
}
```
These can be used to gracefully shutdown instances, which is necessary if a service has long running jobs e.g. a `worker` service.
#### Halt
Asgard supports manually stopping a release while is it being deployed. Just execute:
```
step-asg-deployer halt deploy-test-release.json
```
This will:
1. Find the currently running deploy for the project configuration
2. Write a `halt` file to S3
3. Wait for Asgard to detect the halt file and fail the deploy
<img src="./assets/sad-halt.gif" alt="Asgard deploy" />
Halt does not guarantee that the release will not be deployed, if executed too late the release may still result in success.
**DO NOT** use `Stop execution` of the Asgard step function as it will not clean up resources and leave AWS in a bad state.
### Security
Deployers are critical pieces of infrastructure as they may be used to compromise software they deploy. As such, we take security very seriously around the `step-asg-deployer` and try to answer the following questions:
1. *Authentication*: Who can deploy?
2. *Authorization*: What can be deployed?
3. *Replay* and *Man-in-the-middle (MITM)*: Can some unauthorized person edit or reuse a release to change what is deployed?
4. *Audit*: Who has done what, and when?
#### Authentication
The central authentication mechanisms are the AWS IAM permissions for step functions and S3.
By limiting the `ec2:CreateAutoscalingGroup`, permissions the Asgard function becomes the only way to deploy ASG's. Then limiting permissions to who can call `states:StartExecution` for Asgard limits who can deploy.
Ensuring that Asgard's lambda can only access a single S3 bucket, further limits who can deploy with:
```yaml
{
"Effect": "Allow",
"Action": [
"s3:GetObject*", "s3:PutObject*",
"s3:List*", "s3:DeleteObject*"
],
"Resource": [
"arn:aws:s3:::#{s3_bucket_name}/*",
"arn:aws:s3:::#{s3_bucket_name}"
]
},
{
"Effect": "Deny",
"Action": ["s3:*"],
"NotResource": [
"arn:aws:s3:::#{s3_bucket_name}/*",
"arn:aws:s3:::#{s3_bucket_name}"
]
},
```
Who can execute the step function, and who can upload to S3 are the two permissions that guard who can deploy.
#### Authorization
All resources that can be used in a Asgard deploy must opt-in using tags or paths. Additionally, service resources require specific tags or paths denoting which project/config/service can use them.
Assets uploaded to S3 are in the path `/<ProjectName>/<ConfigName>` so limiting who can `s3:PutObject` to a path can be used to limit what project-configs they can deploy or halt.
#### Replay and MITM
Each release the client generates a release `release_id`, a `created_at` date, and together also uploads the release to S3.
The `step-asg-deployer` will reject any request where the `created_at` date is not recent, or the release sent to the step function and S3 don't match. This means that if a user can invoke the step function, but not upload to S3 (or vice-versa) it is not possible to deploy old or malicious code.
#### Audit
Working out what happened and when is very useful for debugging and security response. Step functions make it easy to see the history of all executions in the AWS console and via API. S3 can log all access to cloud-trail, so collecting from these two sources will show all information about a deploy.
### Continuing Deployment
There is always more to do:
1. Allow LifeCycle Hooks to send to Cloudwatch.
1. Subnet, AMI, life cycle and userdata overrides per service.
1. Check EC2 instance limits and capacity before deploying.
1. Slowly scale instances up rather than all at once, e.g. deploy 1 instance check it is healthy then deploy the rest.
1. Add ELB and Target Group error rates when checking healthy.
1. Custom auto-scaling policy types.
|
2024-11-18T18:26:03.406841+00:00 | 2020-08-17T07:04:03 | f543b0adbe6e721df3a0fb75f4ead4d3a1f30871 | 3 | {
"blob_id": "f543b0adbe6e721df3a0fb75f4ead4d3a1f30871",
"branch_name": "refs/heads/master",
"committer_date": "2020-08-17T07:04:03",
"content_id": "f37089a2428103fb57db69aae0aa6c648ba25c8b",
"detected_licenses": [
"MIT"
],
"directory_id": "83cb9a5c5e36681c1d9ccbfff50bf4b4d14d1ee6",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": "2014-09-12T17:51:34",
"gha_event_created_at": "2022-01-06T22:48:40",
"gha_language": "Python",
"gha_license_id": "MIT",
"github_id": 23970758,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 704,
"license": "MIT",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:55",
"repo_name": "StevenThuriot/sentry-on-heroku",
"revision_date": "2020-08-17T07:04:03",
"revision_id": "3bfb66079a36a08d67d94fbbdd47c19dfdc6bf1c",
"snapshot_id": "03af3ee5c7b462fa8df08cd8bee008a6aa2f4a85",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/StevenThuriot/sentry-on-heroku/3bfb66079a36a08d67d94fbbdd47c19dfdc6bf1c/README.md",
"visit_date": "2022-01-22T18:43:44.344730"
} | 2.8125 | stackv2 | 
================
[Sentry](http://github.com/getsentry/sentry) is a realtime event logging and aggregation platform.
At its core it specializes in monitoring errors and extracting all the information needed to do a proper post-mortem without any of the hassle of the standard user feedback loop.
Basic setup
-----------
Follow the steps below to get Sentry up and running on Heroku:
1. [](https://heroku.com/deploy)
2. Create a user account for yourself::
heroku run "sentry --config=sentry.conf.py createsuperuser"
That's it!
|
2024-11-18T18:26:03.505744+00:00 | 2020-06-18T14:13:50 | 11b1db5466424673b7e92fdae74baebc3530ed63 | 3 | {
"blob_id": "11b1db5466424673b7e92fdae74baebc3530ed63",
"branch_name": "refs/heads/master",
"committer_date": "2020-06-18T14:13:50",
"content_id": "58f8c3e3b4698780771c2b3b0848774c87df101a",
"detected_licenses": [
"MIT"
],
"directory_id": "6404192a488ffb0b50bff7d96dcf5bd21c38e112",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": "2020-06-18T05:15:21",
"gha_event_created_at": "2021-06-22T15:44:52",
"gha_language": "PHP",
"gha_license_id": "MIT",
"github_id": 273147627,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 1286,
"license": "MIT",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:56",
"repo_name": "Guilherme-fagundes/mail",
"revision_date": "2020-06-18T14:13:50",
"revision_id": "b429f23f4cc69165db2c371e6cc5f792ebb4db5e",
"snapshot_id": "06bf7ac3e3e3ca5d4f22233d57013d592c1babc3",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/Guilherme-fagundes/mail/b429f23f4cc69165db2c371e6cc5f792ebb4db5e/README.md",
"visit_date": "2023-06-05T08:41:53.660753"
} | 2.859375 | stackv2 | # Mail
[](https://packagist.org/packages/coffeebreaks/mail)
[](https://packagist.org/packages/coffeebreaks/mail)
[](//packagist.org/packages/coffeebreaks/mail)
[](https://packagist.org/packages/coffeebreaks/mail)
[](https://scrutinizer-ci.com/g/Guilherme-fagundes/mail)
[](https://scrutinizer-ci.com/g/Guilherme-fagundes/mail)
<p>A simple class email send</p>
#### Instalation
```bash
composer require coffeebreaks/mail
```
##### Example
```php
$mail = new \coffeebreaks\Mail();
$mail->config('smtp.your-host.com', 'your-user',
'you-pass', port);
<EMAIL_ADDRESS>'Suport',
<EMAIL_ADDRESS>'Test classe Mail',
"This is a email with <b>PhpMailer</b> and generic class <i>Mail</i>");
if ($mail->getResult()){
echo "Success :)";
}
``` |
2024-11-18T18:26:03.549862+00:00 | 2017-03-04T18:47:59 | 3cac044d8483c70869c6b1e44dd1cbbc879ab8f3 | 4 | {
"blob_id": "3cac044d8483c70869c6b1e44dd1cbbc879ab8f3",
"branch_name": "refs/heads/master",
"committer_date": "2017-03-04T18:47:59",
"content_id": "d683c5235c91bd360eb3016be957e70e44d9807d",
"detected_licenses": [
"Apache-2.0"
],
"directory_id": "bdfc6d7ae465d4526287a793c562c7089b6bc7ee",
"extension": "md",
"filename": "chapter4_1.md",
"fork_events_count": 1,
"gha_created_at": "2017-03-04T22:07:45",
"gha_event_created_at": "2017-03-04T22:07:45",
"gha_language": null,
"gha_license_id": null,
"github_id": 83927816,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 2842,
"license": "Apache-2.0",
"license_type": "permissive",
"path": "/en/part4/chapter4_1.md",
"provenance": "stack-edu-markdown-0000.json.gz:57",
"repo_name": "ivanjunckes/jnosql-book",
"revision_date": "2017-03-04T18:47:59",
"revision_id": "8c0009e715ad4640e212ff901ee9744a0b4f85e1",
"snapshot_id": "cd5930d42dc41889bcd32e7a1d73890948c08ef5",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/ivanjunckes/jnosql-book/8c0009e715ad4640e212ff901ee9744a0b4f85e1/en/part4/chapter4_1.md",
"visit_date": "2021-01-20T09:57:05.276905"
} | 3.609375 | stackv2 | ## Models Annotation
As mentioned previously, Artemis has annotations that make the Java developer life easier; this annotation has two categories:
* Annotations Models
* Qualifier annotation
#### Annotations Models
The annotation Models is to converts the entity model to the entity on communication, the Diana entity:
* Entity
* Column
* MappedSuperclass
* Key
##### Entity
This annotation maps the class to Artemis. It has a unique attribute: name. This attribute is to inform either the column family name or the document collection name, etc.. The default value is the simple name class, for example, given the org.jnosql.demo.Person class the default name will `Person`.
```java
@Entity
public class Person {
}
```
```java
@Entity("name")
public class Person {
}
```
##### Column
This annotation it to define which fields on an Entity will be persisted. It also has a unique attribute name to specify that name on Database, and the default value is the field name.
```java
@Entity
public class Person {
@Column
private String nickname;
@Column
private String name;
@Column
private List<String> phones;
//ignored
private String address;
//getter and setter
}
```
##### MappedSuperclass
If this annotation puts on a Parent class, the Artemis will persist its information as well. So beyond the son class, Artemis will store any field that is in Parent class with Column annotation.
##### Key
Just to Key-value database, that shows on the key-value database with a field is a key.
```java
@Entity
public class User implements Serializable {
@Key
private String userName;
private String name;
private List<String> phones;
}
```
#### Qualifier annotation
That is important to work with more than one type of the same application.
```java
@Inject
private DocumentRepository repositoryA;
@Inject
private DocumentRepository repositoryB;
```
With the same two injections to the same interface, CDI throws an ambiguous exception. There is the `Database` qualifier to fixes problem like this. It has two attributes:
* **DatabaseType**: The database type, key-value, document, column, graph.
* **privider**: The provider database name, eg. "cassandra, "hbase", "mongodb". So using the `Database` qualifier:
```java
@Inject
@Database(value = DatabaseType.DOCUMENT, provider = “databaseA”)
private DocumentRepository repositoryA;
@Inject
@Database(value = DatabaseType.DOCUMENT, provider = “databaseB”)
private DocumentRepository repositoryB;
```
Beyond this annotation, the producer method with the entity manager is required.
The benefits of use this qualifier instead of creating a new one. If the Manager Entity be produced using `Database` as a qualifier, Artemis will create classes such as DocumentRepository, ColumnRepository, etc. automatically.
|
2024-11-18T18:26:03.755759+00:00 | 2023-08-02T11:30:37 | a18d5bd9e942684930255cf8341d4d19cfcc4103 | 4 | {
"blob_id": "a18d5bd9e942684930255cf8341d4d19cfcc4103",
"branch_name": "refs/heads/master",
"committer_date": "2023-08-02T11:30:37",
"content_id": "dc6a41c20657f4d72c27ca25c62cee63ab152716",
"detected_licenses": [
"MIT"
],
"directory_id": "ed024736c98255d4b309a1694dd45f4912d44f7b",
"extension": "md",
"filename": "CONTRIBUTING.md",
"fork_events_count": 342,
"gha_created_at": "2018-10-12T22:41:36",
"gha_event_created_at": "2023-09-13T09:39:01",
"gha_language": "Java",
"gha_license_id": "MIT",
"github_id": 152814648,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 5170,
"license": "MIT",
"license_type": "permissive",
"path": "/CONTRIBUTING.md",
"provenance": "stack-edu-markdown-0000.json.gz:58",
"repo_name": "jenkinsci/warnings-ng-plugin",
"revision_date": "2023-08-02T11:30:37",
"revision_id": "db07bfe4addb69a649ffc1cfcc7ec72ad3314041",
"snapshot_id": "02caee0467eae6f1c1ec662b2f87598fedaaae0f",
"src_encoding": "UTF-8",
"star_events_count": 344,
"url": "https://raw.githubusercontent.com/jenkinsci/warnings-ng-plugin/db07bfe4addb69a649ffc1cfcc7ec72ad3314041/CONTRIBUTING.md",
"visit_date": "2023-08-31T05:35:47.546531"
} | 3.59375 | stackv2 | # Contributing to the Warnings Next Generation Plug-in
This document provides information about contributing code to Jenkins' Warnings Next Generation plug-in.
:exclamation: There's a lot more to the Jenkins project than just code. For information on contributing to the Jenkins'
project overall, check out [Jenkins' contributing landing page](https://jenkins.io/participate/).
## Beginner Topics
If you don't have a specific problem or task in mind, i.e., you simply want to participate in this open source project
I would suggest looking at the
[open newbie friendly issues in our issues tracker](https://issues.jenkins.io/issues/?filter=-1&jql=resolution%20%3D%20Unresolved%20AND%20component%20in%20(analysis-model%2C%20forensics-api-plugin%2C%20git-forensics-plugin%2C%20warnings-ng-plugin)%20AND%20labels%20%3D%20newbie-friendly%20order%20by%20updated%20DESC).
I marked several newbie friendly issues with the label `newbie-friendly`. These are a good starting
point to get in touch with this Jenkins plugin.
If you already have some experience with the plugin you can also fix one of the issues that are marked with the label
`help-wanted` [open issues in our issues tracker](https://issues.jenkins.io/issues/?filter=-1&jql=resolution%20%3D%20Unresolved%20AND%20component%20in%20(analysis-model%2C%20forensics-api-plugin%2C%20git-forensics-plugin%2C%20warnings-ng-plugin)%20AND%20labels%20%3D%20help-wanted%20order%20by%20updated%20DESC).
## Parser Implementations
If you are planning to provide your own parser, please also have a look at the project
[Static Analysis Model and Parsers](https://github.com/jenkinsci/analysis-model). Here, all parsers need to be
added. The Jenkins Warnings Plug-in does not include the parsers anymore, it just links all parsers using the
analysis-model library.
## Getting started
Setup your development environment as described in
[Development environment for Jenkins' Warnings and Code Coverage Plugins](https://github.com/uhafner/warnings-ng-plugin-devenv).
## Coding Guidelines
Start reading the code, and you'll get the hang of it. A complete description of the
coding guidelines is part of a [separate GitHub repository](https://github.com/uhafner/codingstyle) that I am also using
for my lectures about software development.
For [IntelliJ IDEA](https://www.jetbrains.com/idea/) users: If you use the
[provided development environment](https://github.com/uhafner/warnings-ng-plugin-devenv) then the coding style is stored
in configuration files below the `codingstyle/.idea` and `codingstyle/etc` folders. When you import all projects into IntelliJ
then this style will be used automatically.
Moreover, (since this project is about static code analysis :wink:) a configuration for the following static code
analysis tools is defined in the POM (and the `etc` and `.idea` folders of the `codingstyle` module):
- [Checkstyle](http://checkstyle.sourceforge.net/)
- [PMD](https://pmd.github.io/)
- [SpotBugs](https://spotbugs.github.io)
- [Error Prone](http://errorprone.info)
- [IntelliJ](https://www.jetbrains.com/help/idea/code-inspection.html)
This configuration will be picked up automatically if you build the project using Maven. If you install the CheckStyle
plugin of IntelliJ then the correct set of CheckStyle rules will used automatically. Moreover, the code formatter and
the inspection rules will be automatically picked up by IntelliJ.
## Proposing Changes
All proposed changes are submitted and code reviewed using the
[GitHub Pull Request](https://help.github.com/articles/about-pull-requests/) process.
To submit a pull request:
1. Commit changes and push them to your fork on GitHub.
It is a good practice is to create branches instead of pushing to master.
2. In GitHub Web UI click the **New Pull Request** button.
3. Select `warnings-ng-plugin` as **base fork** and `master` as **base**, then click **Create Pull Request**.
4. Fill in the Pull Request description. It should reflect the changes, the reason behind the changes, and if available a
reference to the Jenkins ticket in our [issue tracker](https://issues.jenkins.io/).
5. Click **Create Pull Request**.
6. Wait for CI results and reviews.
7. Process the feedback (see previous step). If there are changes required, commit them in your local branch and push them
again to GitHub. Your pull request will be updated automatically. Review comments for changed lines will become outdated.
Once your Pull Request is ready to be merged, the repository maintainer will integrate it.
There is no additional action required from pull request authors at this point.
## Copyright
The Static Analysis Suite is licensed under [MIT license](./LICENSE). We consider all contributions as MIT unless it's
explicitly stated otherwise. MIT-incompatible code contributions will be rejected.
Contributions under MIT-compatible licenses may be also rejected if they are not ultimately necessary.
## Continuous Integration
The Jenkins project has a Continuous Integration server... powered by Jenkins, of course.
The CI job for this project is located at [ci.jenkins.io](https://ci.jenkins.io/job/Plugins/job/warnings-ng-plugin/).
|
2024-11-18T18:26:03.790767+00:00 | 2019-08-01T16:41:30 | 7104e77300c1089f09663246f9a4af7daa953ad2 | 3 | {
"blob_id": "7104e77300c1089f09663246f9a4af7daa953ad2",
"branch_name": "refs/heads/master",
"committer_date": "2019-08-01T16:41:30",
"content_id": "b87f39a63be084e8d53f7fcfa34c938bd7bb9c36",
"detected_licenses": [
"Apache-2.0"
],
"directory_id": "8ad9911632b2fb5b7ec38cb4e00b42acab9cf785",
"extension": "md",
"filename": "ISSUE_TEMPLATE.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": null,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 935,
"license": "Apache-2.0",
"license_type": "permissive",
"path": "/ISSUE_TEMPLATE.md",
"provenance": "stack-edu-markdown-0000.json.gz:59",
"repo_name": "isabella232/istio-release",
"revision_date": "2019-08-01T16:41:30",
"revision_id": "9d4b0604e1c8cd869ce8a1c50e77ba4b02ac9509",
"snapshot_id": "4ea5c5e0fea592d5c7a2b98b49f291a69e43ddba",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/isabella232/istio-release/9d4b0604e1c8cd869ce8a1c50e77ba4b02ac9509/ISSUE_TEMPLATE.md",
"visit_date": "2022-04-16T08:08:56.567460"
} | 2.96875 | stackv2 | Thanks for submitting an issue to `istio-release`. We are always trying to improve! To help us, please fill out the following template and our bot will add a story to our icebox. From there we will work to prioritize it on the backlog.
## Issue
[provide quick introduction so this issue can be triaged]
## Context
[provide more detailed introduction and the context on what the impact of this is to your operations e.g. this is critical and is impacting all our apps on production]
## Steps to Reproduce
[ordered list the process to finding and recreating the issue, example below]
## Expected result
[describe what you would expect to have resulted from this process]
## Current result
[describe what you currently experience from this process, and thereby explain the bug]
## Possible Fix
[not obligatory, but suggest fixes or reasons for the bug]
## `name of issue` Output Results
[if relevant, include a screenshot]
|
2024-11-18T18:26:03.828032+00:00 | 2018-07-29T16:18:00 | 0a9db4838a49b7bf7d7d0139fb8cf566d8da4952 | 3 | {
"blob_id": "0a9db4838a49b7bf7d7d0139fb8cf566d8da4952",
"branch_name": "refs/heads/master",
"committer_date": "2018-07-29T16:18:00",
"content_id": "bc56f570ea7a6b1a8dcd4b90808da5491e47ec41",
"detected_licenses": [
"MIT"
],
"directory_id": "7aded1fab37f28b8169580655a274272fb05d2f1",
"extension": "md",
"filename": "2015-03-16-march-match-details.md",
"fork_events_count": 1,
"gha_created_at": "2015-02-16T07:13:22",
"gha_event_created_at": "2015-04-28T15:27:38",
"gha_language": "HTML",
"gha_license_id": null,
"github_id": 30857677,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 2345,
"license": "MIT",
"license_type": "permissive",
"path": "/_posts/news/2015-03-16-march-match-details.md",
"provenance": "stack-edu-markdown-0000.json.gz:60",
"repo_name": "leaberry/jekyll-udpl",
"revision_date": "2018-07-29T16:18:00",
"revision_id": "36c35e95e50ab4ddaa0d9633bb9beb9eada56e1f",
"snapshot_id": "ff37b807c99982bf80dcaed8402c24f124aae8c7",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/leaberry/jekyll-udpl/36c35e95e50ab4ddaa0d9633bb9beb9eada56e1f/_posts/news/2015-03-16-march-match-details.md",
"visit_date": "2021-01-17T07:49:40.477189"
} | 2.796875 | stackv2 | ---
title: "March Match Details"
layout: article
categories: Announcements
---
After the Man vs Man match last month we found a missing magazine. If this looks like your magazine please contact [Michael Scott](mailto:[email protected]).

### March 21st 3 Gun match
Round count - 90 Rifle, 90 Pistol, 90 Birdshot, 5-10 slugs. These are minimums assuming no misses so plan accordingly.
There are 8 stages, with two of them being 3 Gun Nation classifiers. There are several optional targets and the longest required rifle shot is 300 yards. A bonus rifle target will be placed at 400 yards.
### March 28th IDPA match
Here is a link to the [4 stages](https://docs.google.com/presentation/d/1h1oeG_ogPfaAf5VgLfEJ1vh0QOKKohcT84DCoTNQ7iA/edit?usp=sharing) for this months IDPA match.
The minimum round count for the 4 defensive stages is 55+ rounds. As always the classifier is 90 rounds minimum. Bring extra for reshoots. We will have the classifier set up in 6 bays to keep things rolling. We will be using an open bay rotation. Squads with members who aren't shooting the classifier are encouraged to shoot the defensive stages first. All others squads are encouraged to find a classifier bay first.
According to IDPA rules, participants in any Tier 2 match or higher match (such as this years Utah State IDPA Championship) must have a current IDPA classification. A classification is considered current if the competitor has either shot a classifier in the last 12 months or has participated in a tier 2 or higher sanctioned match in the last 12 months. Last year's Utah State IDPA Championship was not sanctioned and does not count toward keeping a competitor's classification current. Competitors whose classification will not be current as of the Utah State IDPA match (May 23rd) are encouraged to shoot the classifier this month.
There are about 35 open slots left in the 2015 Utah State IDPA Championship. Participants who have registered but not paid are not guaranteed a slot in the match. Once registration has filled, those who have not paid will be moved to the waiting list so paid registrants can squad in their place. We will try to give adequate warning before this happens. However registration is expected to speed up as the registration deadline gets closer.
|
2024-11-18T18:26:03.875514+00:00 | 2022-10-26T03:09:33 | e43e654ee61cb565f1cf1f6eddbca6f1cbe5d7a4 | 3 | {
"blob_id": "e43e654ee61cb565f1cf1f6eddbca6f1cbe5d7a4",
"branch_name": "refs/heads/master",
"committer_date": "2022-10-26T03:09:33",
"content_id": "6bc522966fcf89a35e90db2ddd79341e10db72de",
"detected_licenses": [
"MIT"
],
"directory_id": "2fe327bb9a415480d775f0a92c25d119adf50ae3",
"extension": "md",
"filename": "2020-07-06-starting-ds-projects.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": 92595943,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 14815,
"license": "MIT",
"license_type": "permissive",
"path": "/_posts/2020-07-06-starting-ds-projects.md",
"provenance": "stack-edu-markdown-0000.json.gz:61",
"repo_name": "Corbin-A/Corbin-A.github.io",
"revision_date": "2022-10-26T03:09:33",
"revision_id": "c5b3582a8d49f53333b4752c8aaf9c96d80a3912",
"snapshot_id": "4b05c1cba56a01441b2942494bdbe6b6660cbb33",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/Corbin-A/Corbin-A.github.io/c5b3582a8d49f53333b4752c8aaf9c96d80a3912/_posts/2020-07-06-starting-ds-projects.md",
"visit_date": "2022-11-04T01:51:21.603236"
} | 3.390625 | stackv2 | ---
layout: post
title: Starting a Data Science Project from Scratch
---
For those that don't know, I am a Senior Data Scientist at IBM (specifically in GBS, IBM's consulting arm). As a Sr. DS, I am tasked with leading small teams of data scientists to delivery on client engagements which are generally around 6 - 12 weeks long. As such, I have had to start quite a few distinct DS projects. Surprisingly, there is no established set of best practices for technical leads to follow within IBM, so I have had to hone my own process over the years. Since I've had to learn lots of things the hard way, I thought I would write down my process here for my own reference and for anyone else to take and use themselves.
With that said, I want to mention some caveats:
1. There are almost certainly better ways to do certain things that I don't know about. Please leave suggestions in the comments!
2. Every project is different. This is a rough guide for myself and others and should be treated as such.
3. I code almost exclusively in python for Data Science. I do not know the R ecosystem well enough to give advice on R projects.
4. My expertise is in data science, which has a distinctly different project structure from traditional SE. There will be some crossover in this list, but it will be tailored towards data science (hell, it's in the title).
This page will deal specifically with creating:
1. The overall directory structure
2. An `environment.yml` file seeded with necessary packages
3. A `setup.py` file for easy `pip install -e .`
4. A `.pre-commit-config.yaml` file to enforce nbstripout, flake8 and mypy via pre-commit git hook
5. `tests/` directory and notes around testing in DS projects
6. A `.travis.yml` or other CI recipe file
7. Data infrastructure
8. Getting Started section in README
### Directory structure
This should be flexible to accomodate the needs of the project, but a great starting point can be found at [cookiecutter-datascience](https://github.com/drivendata/cookiecutter-data-science). When I personally start off a project, I remove some of the pre-populated files and get down to the following directory structure:
```
├── .gitignore
│
├── README.md <- The top-level README for developers using this project.
│
├── data
│ ├── external <- Data from third party sources.
│ ├── interim <- Intermediate data that has been transformed.
│ ├── processed <- The final, canonical data sets for modeling.
│ └── raw <- The original, immutable data dump.
│
├── docs <- A default Sphinx project; see sphinx-doc.org for details
│
├── models <- Trained and serialized models, model predictions, or model summaries
│
├── notebooks <- Jupyter notebooks
│
├── references <- Data dictionaries, manuals, and all other explanatory materials.
│
├── reports <- Generated analysis as HTML, PDF, LaTeX, etc.
│ └── figures <- Generated graphics and figures to be used in reporting
│
├── tests <- Automated pytest tests
│
└── <pkg_name> <- Source code for use in this project.
├── __init__.py <- Makes src a Python module
│
├── data_prep <- Scripts to download or generate data
│ └── make_dataset.py
│
├── feat_eng <- Scripts to turn raw data into features for modeling
│ └── build_features.py
│
├── models <- Scripts to train models and then use trained models to make
│ │ predictions
│ ├── predict_model.py
│ └── train_model.py
│
└── visualization <- Scripts to create exploratory and results oriented visualizations
└── visualize.py
```
This gets us started, but we need to set up a bit more infrastructure.
### Write minimal `setup.py` file
When your team starts writing `.py` files, they will need to be importing from around the `<pkg_name>` directory. Imports are a major PITA in python if you don't locally pip install the project. To allow for absolute imports, you should create a minimal `setup.py` file such as this:
```python
from setuptools import setup, find_packages
setup(
name='name-of-project',
version='0.1.0',
description='short description of project',
author='author name',
<EMAIL_ADDRESS> url='github pages or github link, or whatever you are using for your project',
packages=find_packages(),
license='license-type',
)
```
After this is done, all a dev will need to do is navigate to the project root directory via a terminal and run `pip install -e .`, which will use pip to download the package to their local env and allow for absolute imports in their source code.
### Create intial `environment.yml` or `requirements.txt` file
One difficulty of working on a project with many people is getting everyone on the same page as regards project dependencies. The natural solution is to use a virtual environment and upload a `requirements.txt` or `environment.yml` file to github. For DS projects, I recommend using conda environments, as their dependency management is (supposedly) miles above the competition as regards dependency management for scientific libraries (though my experience is admittedly limited on some of the other options like `poetry`).
In addition to this, there are some libraries that are so ubiquitous in Data Science projects that I recommend seeding the `environment.yml` file with some of these top libraries so that everyone starts on the same page. These packages can include numpy, pandas, sklearn, and jupyter, as well as linting, testing, and misc libraries such as mypy, black, flake8, pytest, nbstripout, pre-commit, etc. Depending on the project, XGBoost and fbprophet may also fit the bill.
(NOTE: If your project involves time series and you plan on using fbprophet, it currently does not support python 3.8 as of the time of this writing, so make sure to either include `fbprophet` in your list of seeded libraries, or specify `python=3.7` when you create your conda env.)
So my recommendation is to run the following commands in the project root directory:
```bash
conda create -n <shorthand-project-name-for-env> numpy pandas scikit-learn jupyter jupyter_contrib_nbextensions mypy flake8 pytest pytest-cov nbstripout pre-commit black
conda env export --no-builds | grep -v "prefix" > environment.yml
```
The second command will export the environment with version numbers (but not build identifiers) to a file called `environment.yml` which devs can then use by navigating to the project root directory and running `conda env update && conda activate <shorthand-project-name-for-env>`. This will ensure that all devs are using conda, virtual environments, and everyone is using the same versions of the most important packages. As the project evolves and dependencies grow, you may need to consolidate various envs, but this gets all devs started off on the same page.
### Create `.pre-commit-hooks.yaml` file
For the uninitiated, git provides a miniature version of Continuous Integration (CI) via git hooks. I recommend looking these up, but essentially you can instruct git to run certain code at various stages of the git workflow, noteably `pre-commit`. What this means is you can force certain linting standards to be met before devs can commit their code and consequently push upstream to their branches. Everyone is going to have a different opinion about this, but I think the following checks are highly beneficial to include as pre-commit hooks
- [nbstripout](https://github.com/kynan/nbstripout): No data science project that utilizes notebooks should be without this program as a pre-commit hook. `nbstripout` ensures that all output has been stripped from notebooks which is very beneficial for VC. See [this video](https://www.youtube.com/watch?v=BEMP4xacrVc) for evidence of how useful this is.
- mypy: I am firmly of the belief providing type hints in python makes for better code, so I recommend enforcing their use
- black: This is an auto-formatter for python. Spend less time learning the individual coding styles of each of your developers and instead let black enforce standardized best practices.
- flake8: Helps adhere to PEP8 standards. Feel free to add in any ignores that better suit your company's style guide
Here is an example `.pre-commit-hooks.yaml` file that I use for all of my projects (note that you will need to edit one of the mypy entries):
```
repos:
- repo: https://github.com/pre-commit/pre-commit-hooks
rev: v2.4.0
hooks:
- id: check-yaml
- id: end-of-file-fixer
- id: trailing-whitespace
- repo: https://github.com/psf/black
rev: stable
hooks:
- id: black
language_version: python3.7
- repo: https://github.com/kynan/nbstripout
rev: master
hooks:
- id: nbstripout
files: ".ipynb"
- repo: https://gitlab.com/pycqa/flake8
rev: 3.7.8
hooks:
- id: flake8
types: [file, python]
args: ['--ignore=E203,E266,E401,E501,W503,F403,F401', '--max-line-length=120']
- repo: https://github.com/pre-commit/mirrors-mypy
rev: v0.780
hooks:
- id: mypy
args: [--ignore-missing-imports, --disallow-untyped-defs]
files: notebooks/
- id: mypy
args: [--ignore-missing-imports, --disallow-untyped-defs]
files: <pkg_name>
```
One thing I will note here is that I do not include pytest in my pre-commit hooks. This is because testing can take a lot of time and can consequently discourage devs from commiting often (as they should be doing). However, pytest is included in the CI pipeline, which is discussed later in the article. However, speaking of testing...
### Create `tests/` directory
Code should be tested, and DS code is no different. I very rarely see this discussed, further evidenced by the fact that there is no tests directory in the cookiecutter project template which baffles me. If we are scientists, what is more important than ensuring accurate code and reproducibility? Tests help us do this, and instantiating a culture of testing in your team will make life a lot easier in the long run. In data science there are a lot of assumptions that are made, such as "I don't expect there to be nans in this column" or "this is a cumulative column, so it should be monotonically increasing". However, these assumptions do not always hold up over time (especially if another team manages the underlying data) and code / models can very easily break if underlying assumptions are broken. Therefore we should all be baking tests into our workflows and ensuring our assumptions stay valid and our code covers edge cases where appropriate.
For testing in python I recommend using pytest. If you're not already familiar, there are endless resources online. I can recommend the following video: [link](https://www.youtube.com/watch?v=4fUzlBbLOaw). You can also incorporate propert-based testing using the [`hypothesis`](https://hypothesis.readthedocs.io/en/latest/) library to generate tests which the linked video also covers.
### Set up CI
Figure out what service your company uses for continuous integration and make the appropriate file for integrating CI into your project. For instance, IBM uses TravisCI, so I use the following file in for my CI pipeline:
```
branches:
only:
- master
language: python
python:
- "3.6"
install:
- sudo apt update
# We do this conditionally because it saves us some downloading if the
# version is the same.
- wget https://repo.continuum.io/miniconda/Miniconda3-latest-Linux-x86_64.sh -O miniconda.sh;
- bash miniconda.sh -b -p $HOME/miniconda
- source "$HOME/miniconda/etc/profile.d/conda.sh"
- hash -r
- conda config --set always_yes yes --set changeps1 no
- conda config --add channels conda-forge
- conda config --set channel_priority strict
- conda update -q conda
# Useful for debugging any issues with conda
- conda info -a
- conda env update
- conda activate <env_name>
- python setup.py install
script:
- python -m pytest tests --cov=<pkg_name> --cov-fail-under=<desired_coverage_percentage>
- mypy --ignore-missing-imports --disallow-untyped-defs <pkg_name>
- flake8 --ignore=E203,E266,E401,E501,W503,F403,F401 --max-line-length=120 <pkg_name>
```
The mypy and flake8 checks will already be handled by the pre-commit git hooks we mentioned earlier, but this will ensure that the devs adhered to the reqs. NOTE: You will need to change the `--cov=`, mypy and flake8 options/args to your package name. Additionally, change the `--cov-fail-under` value to your intended coverage percentage.
### Set up Data Infrastructure
The next thing on your list of to-do's should be to figure out what your data infrastructure is going to be. This _could_ be flat files in the data directory of the project, but I really don't recommend this as a longterm solution. What you should implement will depend on the project (as is the case for everything in this post). If your data is large, you might already have a hadoop ecosystem in place. If you're using more manageable datasets, why not use a tried and true RDBMS? Prefer NoSQL? That's fine too. One thing that I still need to do some more research into is DVC (Data Version Control), but it seems very promising.
### Write "Getting Started" section of README file
Make it as easy as possible for your devs to get up and running now that you have implemented the above structures. An example "Getting Started" section may look like the below:
```
### Getting Started
This project assumes you have Anaconda or Miniconda installed on your machine. If you do not, please install from https://docs.conda.io/en/latest/miniconda.html
1. `git clone` this repo in the desired directory on your local machine
2. `cd` into the project directory
3. Run `conda env update && conda activate <env_name>`
4. Run `pip install -e .`
5. Run `pre-commit install`
```
Note: If you plan on using Sphinx for documentation automation, then make sure your READMEs are in RST instead of the more traditional markdown.
## Room for improvement
There's lots. I'm still fleshing out some of these files and processes, but some immediate room for improvement is the following:
- Separate test-environment.yaml file that does not install jupyter and its dependencies in your CI system. This is a very heavy package and makes CI take longer than necessary.
- Notes around documentation automation via sphinx or pdoc
- Notes around project management via github issues & github projects (trello, jira, etc. as alternatives)
- Research DVC and have a better solution for data version management.
|
2024-11-18T18:26:03.932896+00:00 | 2017-09-08T01:46:06 | 30a51c2a6656baee55f74004e0e5cfc4f058354c | 3 | {
"blob_id": "30a51c2a6656baee55f74004e0e5cfc4f058354c",
"branch_name": "refs/heads/master",
"committer_date": "2017-09-08T01:46:06",
"content_id": "22d9bca0c5a88d0b5d9696ce56501e6749a95a10",
"detected_licenses": [
"MIT"
],
"directory_id": "28cc4444b81667304f378829c8cb1278b7100325",
"extension": "md",
"filename": "README.md",
"fork_events_count": 1,
"gha_created_at": "2016-10-27T08:10:51",
"gha_event_created_at": "2017-09-08T01:46:08",
"gha_language": "Ruby",
"gha_license_id": null,
"github_id": 72085908,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 880,
"license": "MIT",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:62",
"repo_name": "yengfinity/json_editor_rails",
"revision_date": "2017-09-08T01:46:06",
"revision_id": "c16ac4f7baa73d188000ec08f5c12a73074d17e4",
"snapshot_id": "39ca86365fbf5ee54ee4dfbad4ddff0aba908b7a",
"src_encoding": "UTF-8",
"star_events_count": 2,
"url": "https://raw.githubusercontent.com/yengfinity/json_editor_rails/c16ac4f7baa73d188000ec08f5c12a73074d17e4/README.md",
"visit_date": "2021-01-18T19:33:54.648963"
} | 3.046875 | stackv2 | # Json Editor Ruby on Rails integration gem
All credits go to Jeremy [and his Javascript Library ](https://github.com/jdorn/json-editor)
## Installation
Add this line to your application's Gemfile:
gem 'json_editor_rails'
And then execute:
$ bundle
Or install it yourself as:
$ gem install json_editor_rails
Now you need to edit your app/assets/javascripts/application.js file and add the following line:
``` javascript
//= require jsoneditor
```
## Usage
Jeremy has some great info on this: [Library usage](https://github.com/jdorn/json-editor#usage)
## Contributing
1. Fork it ( https://github.com/sharshenov/json_editor_rails-rails/fork )
2. Create your feature branch (`git checkout -b my-new-feature`)
3. Commit your changes (`git commit -am 'Add some feature'`)
4. Push to the branch (`git push origin my-new-feature`)
5. Create a new Pull Request |
2024-11-18T18:26:03.989078+00:00 | 2023-08-11T03:31:01 | 4daadbebe092d4f79d6ba77fd8e8fac8955da516 | 3 | {
"blob_id": "4daadbebe092d4f79d6ba77fd8e8fac8955da516",
"branch_name": "refs/heads/master",
"committer_date": "2023-08-11T03:31:01",
"content_id": "2c8f5db9b8c521a4b09af7baba49535f85188b6b",
"detected_licenses": [
"BSD-3-Clause"
],
"directory_id": "acefbe29a8fb244c19a610381f0d826a08d939e6",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": 233113707,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 1002,
"license": "BSD-3-Clause",
"license_type": "permissive",
"path": "/v0160gcc/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:63",
"repo_name": "SCOREC/dcs-spack-config",
"revision_date": "2023-08-11T03:31:01",
"revision_id": "bcd0f9c43ae474c0bc7d9d4c8df93b6caea13f14",
"snapshot_id": "60b036979e3c0e77f7f7e35ed6c08f9b5808f5e5",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/SCOREC/dcs-spack-config/bcd0f9c43ae474c0bc7d9d4c8df93b6caea13f14/v0160gcc/README.md",
"visit_date": "2023-09-04T04:29:54.743295"
} | 2.9375 | stackv2 | # dcs-spack-config
CCI DCS (AiMOS) spack configuration and scripts for building the XGC depdencies
with the GCC compilers and OpenMPI.
## contents
config.yaml - global config
modules.yaml - hierarchical layout for lua modules
README.md - this file
setupSpack.sh - env needed for executing spack commands
spack.yaml - spack stacks/environment specification
## setup
```
git clone [email protected]:spack/spack.git spack
cd !$
git checkout v0.16.0
# cp setupSpack.sh and edit the path to the spack repo
source setupSpack.sh
# create the environment
spack env create v0160
spack env activate v0160
# add gcc8.4 and find existing packages - already included in spack.yaml
# spack compiler find /gpfs/u/software/ppc64le-rhel7/gcc/8.4.0/1/bin
# spack external find
# copy the yaml files into the v0160
cp /path/to/this/dir/*.yaml var/spack/environments/v0160/.
```
## resuming work in an environment
```
cd /gpfs/u/software/dcs-spack-src/spack_v0160
source setupSpack.sh
spack env activate v0160
```
|
2024-11-18T18:27:53.269733+00:00 | 2023-02-03T14:19:17 | dcce3e353ceb1d475a315ed5e6a957c5adac4efd | 3 | {
"blob_id": "dcce3e353ceb1d475a315ed5e6a957c5adac4efd",
"branch_name": "refs/heads/master",
"committer_date": "2023-02-03T14:19:17",
"content_id": "3693b4dbe0ebbdba97719d6f100b0be54b5b2808",
"detected_licenses": [
"MIT"
],
"directory_id": "327f40f40542f02f60d5505033a368967ee1b32f",
"extension": "md",
"filename": "2018-09-24-tutorial-JN-win10.md",
"fork_events_count": 2,
"gha_created_at": "2018-09-24T20:33:34",
"gha_event_created_at": "2022-10-06T02:25:56",
"gha_language": "JavaScript",
"gha_license_id": "MIT",
"github_id": 150165003,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 1826,
"license": "MIT",
"license_type": "permissive",
"path": "/_posts/2018-09-24-tutorial-JN-win10.md",
"provenance": "stack-edu-markdown-0000.json.gz:64",
"repo_name": "qMRLab/qmrlab.github.io",
"revision_date": "2023-02-03T14:19:17",
"revision_id": "028a7a6d43e772c29de89126ae8f366d5d4bf78d",
"snapshot_id": "647bd3d06fd37f8806b7234c7de8c5046f9a99d8",
"src_encoding": "UTF-8",
"star_events_count": 1,
"url": "https://raw.githubusercontent.com/qMRLab/qmrlab.github.io/028a7a6d43e772c29de89126ae8f366d5d4bf78d/_posts/2018-09-24-tutorial-JN-win10.md",
"visit_date": "2023-02-04T22:17:21.281070"
} | 3.296875 | stackv2 | ---
layout: post
title: "qMRLab in a Jupyter Notebook: A tutorial for Win10 users"
author: Agah
categories: [ jekyll ]
image: /assets/images/tut1.png
feature-image: assets/images/tut1.png
featured: true
hidden: false
---
Imagine that you are exploring a new qMRI analysis method. The code is open in an editor window and your command window is swarmed with hundreds of lines of output logs and command histories. At the same time, you are browsing some paper or documentation to understand how that method works. If you have ever wished for a single piece of document that explains the theory and allows you to play around with the code, you may already have discovered [Jupyter Notebooks](http://jupyter.org/).
Jupyter Notebooks allow you to choose one from over 40 programming languages - including MATLAB and Octave - interactively run your analyses, and share them with other people. Moreover, your code can also create rich, interactive output such as HTML, images, videos, and even LaTeX.
In this first qMRLab tutorial, we walk Windows 10 users through the necessary steps to use Jupyter tools on their local machines for creating lively qMRI notebooks using qMRLab. The tutorial is a [slides.com](http://slides.com) presentation embedded below.
On-screen display mode: To navigate between the slides, please use the blue arrows located at the lower right of the presentation panel.
Full-screen display mode: You can start the full screen mode by clicking the expand icon located at the slides.com toolbar below the presentation. Blue navigation arrows or the direction keys on your keyboard can be used to navigate between slides.
<iframe src="//slides.com/agahkarakuzu/qmrlabjn1/embed" width="900" height="656" scrolling="no" frameborder="0" webkitallowfullscreen mozallowfullscreen allowfullscreen></iframe>
|
2024-11-18T18:27:53.317028+00:00 | 2023-06-11T23:56:03 | 08242c9905e5aa75275e91bcf4362f9d2dce2857 | 4 | {
"blob_id": "08242c9905e5aa75275e91bcf4362f9d2dce2857",
"branch_name": "refs/heads/main",
"committer_date": "2023-06-11T23:56:03",
"content_id": "959067bc20bec85c561d1a370f5c78b9caa91e4e",
"detected_licenses": [
"MIT"
],
"directory_id": "462347e25feba9ad249418a47d567a0a867943a2",
"extension": "md",
"filename": "atom.md",
"fork_events_count": 7,
"gha_created_at": "2016-09-13T19:05:43",
"gha_event_created_at": "2023-03-23T21:35:18",
"gha_language": "Shell",
"gha_license_id": "MIT",
"github_id": 68138648,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 2625,
"license": "MIT",
"license_type": "permissive",
"path": "/topics/atom.md",
"provenance": "stack-edu-markdown-0000.json.gz:65",
"repo_name": "ucsb-cs16/ucsb-cs16.github.io",
"revision_date": "2023-06-11T23:56:03",
"revision_id": "425b1134c04eb1a9074784bcd03754e426f3f4ca",
"snapshot_id": "937b3821a6108b0adb2499bb8f1f156ee67b42cc",
"src_encoding": "UTF-8",
"star_events_count": 6,
"url": "https://raw.githubusercontent.com/ucsb-cs16/ucsb-cs16.github.io/425b1134c04eb1a9074784bcd03754e426f3f4ca/topics/atom.md",
"visit_date": "2023-06-22T18:27:04.128540"
} | 3.71875 | stackv2 | ---
layout: default
parent: Topics
title: "Atom"
description: "graphical editor available on CSIL, plus Mac, Windows, Linux"
---
Atom is a graphical editor that is similar to Sublime Text, but free and open source.
# On machines physically in CSIL and Phelps 3525:
(NOTE: DOES NOT WORK WHEN USING ssh)
To edit hello.cpp with atom on CSIL, type:
```
atom hello.cpp
```
What should happen is that the Unix command prompt returns immediately, but that after a few seconds, a window pops up with the
atom editor
# What if no window pops up?
Are you are probably not logged into a machine directly in CSIL or Phelps 3525? If not: we told you so.
In that case, see the options for using atom remotely, listed next.
# Remote Options
Visual Studio Code always works best when it is run directly on the machine at which you are physically sitting. That is:
* If you are at a machine in Phelps 3525, or the CSIL lab, and you run `atom` directly on that machine, all will be well (typically)
* If you are sitting at your own Windows or Mac machine, or ssh'd in from your own Linux machine, your best bet is to install `atom`
locally on your machine, and edit there.
# But how do I access my file and the compilers on CSIL in that case?
Good question. There several options.
* The most straightforward, but possibly error prone option, is to just copy files back and forth. There are various
[file copy programs](/topics/csil_copying_files/)
* A second option is to use a github repo to move the files back and forth between your local system, and CSIL. This will give you lots
of practice with doing `git add`, `git commit`, `git push`, and `git pull`. However, it has the following down side: it
means that you will need to do a commit with every edit. That is typically not the best practice. We typically want to
do a commit with each unit of meaningful work, i.e. when you've added a a feature, fixed a bug, etc. Doing it with every
edit that you want to test will have you committing work a bit too often. That might not be the worst thing; new users
of git tend to err on the side of having too *few* commits. But still, it isn't the best practice.
* The best way, if you can get it to work, is something called `sshfs`. This is a way of literally mounting your CSIL home
directory so that it *appears* to be local hard drive on your own computer. That is super useful, but deserves
a page of its own, which you can find here: [sshfs](/topics/sshfs/)
# Installing code on Windows, Mac, Linux
For download and installation instructions, visit:
* <https://atom.io/>
|
2024-11-18T18:27:53.378135+00:00 | 2019-01-18T04:23:23 | 90b309c03f73fc1d61ac9f29b21f6e517ce867f9 | 3 | {
"blob_id": "90b309c03f73fc1d61ac9f29b21f6e517ce867f9",
"branch_name": "refs/heads/master",
"committer_date": "2019-01-18T04:23:23",
"content_id": "4a7a5a8d44cbbd9693538e120146fa9312e95694",
"detected_licenses": [
"MIT"
],
"directory_id": "00311d5524afb6ccc9a77084a2f620262e6c6436",
"extension": "md",
"filename": "_index.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": 143807680,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 880,
"license": "MIT",
"license_type": "permissive",
"path": "/content/data_analysis/_index.md",
"provenance": "stack-edu-markdown-0000.json.gz:66",
"repo_name": "GPUE-group/GPUE-group.github.io",
"revision_date": "2019-01-18T04:23:23",
"revision_id": "6bce19b99823f66e2febf24d9349adf127a81ea5",
"snapshot_id": "31d5c9dc843c1454dd7446db6e2b5d3ef26d67ea",
"src_encoding": "UTF-8",
"star_events_count": 1,
"url": "https://raw.githubusercontent.com/GPUE-group/GPUE-group.github.io/6bce19b99823f66e2febf24d9349adf127a81ea5/content/data_analysis/_index.md",
"visit_date": "2020-03-25T13:04:03.780766"
} | 2.59375 | stackv2 | +++
title = "Data analysis"
description = ""
weight = 4
alwaysopen = true
+++
GPUE outputs data in a basic ASCII format so it is easy for users to read the data into an auxiliary program and analyze it as necessary; however, GPUE also provides a series of scripts for 2 and 3D analysis with the following functionality:
1. Plotting in 2 dimensions
2. Generation of 2 dimensional slices of 3 dimensional data
3. Generation of `.vbox` or `.vtk` files in 3 dimensional for plotting in blender and paraview, respectively
4. Generation of images that can later be concatenated into a video with standard tools (`ffmpeg`, `ImageMagick`, etc.)
GPUE is primarily a simulation program and thus provides only limited tools necessary to visualize the data.
If the user requires more advanced data analysis, these must be further developed by the user for their specific research purpose.
|
2024-11-18T18:31:53.151702+00:00 | 2013-10-26T09:14:14 | e99a5985f0f66b62f26f642959fd7538354e4a21 | 2 | {
"blob_id": "e99a5985f0f66b62f26f642959fd7538354e4a21",
"branch_name": "refs/heads/master",
"committer_date": "2013-10-26T09:14:14",
"content_id": "b59f8234e19113a40484083977786db2e6be563e",
"detected_licenses": [
"Apache-2.0"
],
"directory_id": "ce5f97348d1f97d534e5f1a21b9c1e7d11709ded",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": 13499597,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 234,
"license": "Apache-2.0",
"license_type": "permissive",
"path": "/life/Fungi/Basidiomycota/Agaricomycetes/Agaricales/Agaricaceae/Endoptychum/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:67",
"repo_name": "mdoering/backbone",
"revision_date": "2013-10-26T09:14:14",
"revision_id": "712cec6c481c3d78c52698a10644af9aaf7e241d",
"snapshot_id": "67abcd314c8426f4d7c20dbb1a07fca77eec0a97",
"src_encoding": "UTF-8",
"star_events_count": 1,
"url": "https://raw.githubusercontent.com/mdoering/backbone/712cec6c481c3d78c52698a10644af9aaf7e241d/life/Fungi/Basidiomycota/Agaricomycetes/Agaricales/Agaricaceae/Endoptychum/README.md",
"visit_date": "2021-01-19T03:39:00.475878"
} | 2.015625 | stackv2 | # Endoptychum Czern. GENUS
#### Status
ACCEPTED
#### According to
The Catalogue of Life, 3rd January 2011
#### Published in
Bull. Soc. Imp. nat. Moscou 18(2, III): 146 (1845)
#### Original name
Endoptychum Czern.
### Remarks
null |
2024-11-18T18:31:53.280192+00:00 | 2017-07-03T20:57:06 | 831e03b72cdfd7c711c98b418673edb4327a99a2 | 3 | {
"blob_id": "831e03b72cdfd7c711c98b418673edb4327a99a2",
"branch_name": "refs/heads/master",
"committer_date": "2017-07-03T20:57:06",
"content_id": "24ddef691332e741cf6289226b09e5b3e83adafc",
"detected_licenses": [
"Apache-2.0"
],
"directory_id": "93205a7cc036732b95b4002bfc766c02fc2a26e4",
"extension": "md",
"filename": "readme.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": null,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 2789,
"license": "Apache-2.0",
"license_type": "permissive",
"path": "/doc/intro/readme.md",
"provenance": "stack-edu-markdown-0000.json.gz:68",
"repo_name": "FrankGaLu/uctf",
"revision_date": "2017-07-03T20:57:06",
"revision_id": "f7d597b9532995b1509cc29aed4ac7115c2b0cb8",
"snapshot_id": "2cd4e65d481e0f494ea0671ecd5c73277ddc513c",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/FrankGaLu/uctf/f7d597b9532995b1509cc29aed4ac7115c2b0cb8/doc/intro/readme.md",
"visit_date": "2023-03-22T12:03:19.303123"
} | 3.421875 | stackv2 | # Intro
The current concept of a two-swarm game is approximated as a cross between aerial laser tag and aerial capture the flag.
## Scenario
Each swarm controls a set of quadcopters and fixed wings aerial vehicles.
Each aerial vehicle is using the PX4 flight controller.
The vehicles start from the opposite sides of a *game cube* with an edge length of 500 m.
## Rules and Scoring
There are three primary way of scoring points:
* successfully launching and flying for a duration,
* successfully landing within a designated landing zone,
* and successfully tagging aircraft in the other swarm
In the absence of actual lasers and laser detectors or another tagging technology, tagging can be done virtually by transmitting "fire" messages over the network, and having a virtual referee evaluate those fires in light of broadcast pose data from both swarms.
For information about the rules and scoring please see the [arbiter documentation]().
## Real world vs. simulation
The scenario can either be run in the real world or within the Gazebo simulator.
In Gazebo the quadcopter is represented by a *3DR Iris* and the fixed wing by a scaled down *Cessna*.
## Overview of the system architecture
The system architecture is very similar for the real world as well as the simulator.
The following diagram shows all the components involved and how they are connected with each other.

The orange lines depict MAVLINK communication.
Each vehicle uses four ports.
For each connection the port number is indicated by an offset from the vehicle specific base port.
The blue lines depict ROS communication.
Each vehicle uses a namespace constructed from the vehicle type and its unique mavlink system id.
## Coordinate systems
In Gazebo the coordinate system used for the world and models is a follows:
* The `x` direction points from the blue team area towards the gold team area.
* The `y` direction points from the side without any area towards the penalty area.
* Therefore the `z` direction is upwards.
* The coordinate `0, 0, 50` is where the blue team area, the "no" area, and the game cube meet.
The global position information are:
* The `latitude` direction points from the blue team area towards the gold team area.
* The `longitude` direction points from the penalty area towards the side without any area.
The rendered scene in `rqt` (after rotating the view by `pi`) has the following axes:
* The `x` direction points from the penalty area towards the side without any area.
* The `y` direction points from the blue team area towards the gold team area.
* The coordinate `0, 0` is where the blue team area, the penalty area, and the game cube meet.
---
Next: [Setup your system](../setup/readme.md)
|
2024-11-18T18:31:53.309519+00:00 | 2018-02-24T09:06:47 | b6cc8c92aa506f6143d69d7b765cceb1653af2f3 | 3 | {
"blob_id": "b6cc8c92aa506f6143d69d7b765cceb1653af2f3",
"branch_name": "refs/heads/master",
"committer_date": "2018-02-24T09:06:47",
"content_id": "704b141f8f5a824cc34a41c67ddf998dbb4355ce",
"detected_licenses": [
"MIT"
],
"directory_id": "67251df47de04778d5df5229c7ad2fff902444e2",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": 117718503,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 1947,
"license": "MIT",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:69",
"repo_name": "Skchoudhary/Goals",
"revision_date": "2018-02-24T09:06:47",
"revision_id": "95cec63bcaad6a9d698b79eb592d47c98587314d",
"snapshot_id": "58d487359611910934833abc01374aa192a9a560",
"src_encoding": "UTF-8",
"star_events_count": 1,
"url": "https://raw.githubusercontent.com/Skchoudhary/Goals/95cec63bcaad6a9d698b79eb592d47c98587314d/README.md",
"visit_date": "2021-09-07T14:51:41.944903"
} | 3.171875 | stackv2 | # Goals
* To become Good Full Stack Developer.
# Priority Order
1. JavaScrpit[Front-end Developer]
2. Python[Back-end]
3. Django ORM[Back-end]
4. Flexbox CSS[Front-end]
5. HTML5 Canvas API[Front-end]
6. SQl[Databse]
7. MongoDB[Database]
***
## Front-end Developer
* **JavaScrpit**: Goal is to learn semantics, best practice in the language.Want to learn React/AngularJS FrameWork.
* **Flexbox CSS layout**: Want to learn the use of Flexbox for responsive layout.
* **HTML5 Canvas API**: Want to learn implementation and use of the API.
[Also want to learn how to do Browser rendering optimization]
## Back-end Developer
* **Python**: Want to explore more deeper in the Language, best practice and to explore Object Oriented side of the language.As i still consider myself new to the language.Also want to explore Django Framework.
* **Django ORM**: Want to master Django ORM and different tools to effectively communicate with the database.
## Database
* **MongoDB**: Want to learn how to setup and maintain MongoDB database.
* **SQL**: Want to learn Advance best practices to handle database with large dataset. Techniques and way to handel Distributed Datbase.
***
# Short term Task.
## Books
* To read 6 books in a Year.
## Courses.
* To complete these [Core Maths](https://github.com/ossu/computer-science#core-math)
listed courses quarterly goal.
# Random
* To try my hands on Raspberry Pi.
***
# Task to achieve Goal.
## JavaScript
* [You-Dont-Know-JS] (https://github.com/getify/You-Dont-Know-JS)--Time 7 weeks
* [Mozilla JS Guide] (https://developer.mozilla.org/en-US/docs/Web/JavaScript/Guide)--Time 4 weeks
* Mastering React [eBook]--Time 4 weeks
## Python
* [pymbook](http://pymbook.readthedocs.io/en/latest/)--Time 1 week
* Learning Python Design Patterns [eBook] --Time 4 weeks
* Python Unlocked [eBook] --Time 3 Weeks
**CLOCKED TIME FORMAT**
CLOCK: [2017-04-21 Fri 20:15]--[2017-04-21 Fri 22:20] => 2:05
|
2024-11-18T18:31:53.430104+00:00 | 2023-07-30T04:01:16 | 0e346066a61a6e88338bf206522c9454c4cd7cc7 | 3 | {
"blob_id": "0e346066a61a6e88338bf206522c9454c4cd7cc7",
"branch_name": "refs/heads/master",
"committer_date": "2023-07-30T04:01:16",
"content_id": "3a154e82936f20fd6f410edd30d2e631ddf4b70a",
"detected_licenses": [
"MIT"
],
"directory_id": "d939455f557cd9eda2619fd812025d8e19b083e3",
"extension": "md",
"filename": "21. Merge Two Sorted Lists(Easy).md",
"fork_events_count": 20,
"gha_created_at": "2016-05-12T07:52:04",
"gha_event_created_at": "2023-05-25T19:00:48",
"gha_language": "Java",
"gha_license_id": "MIT",
"github_id": 58620185,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 2884,
"license": "MIT",
"license_type": "permissive",
"path": "/Solutions/English/21. Merge Two Sorted Lists(Easy).md",
"provenance": "stack-edu-markdown-0000.json.gz:70",
"repo_name": "longluo/leetcode",
"revision_date": "2023-07-30T04:01:16",
"revision_id": "5a171f223c03cfdddb18488fd4bc5910039e21c4",
"snapshot_id": "61d445bd4fbae13b99ea24e9ef465bb700022866",
"src_encoding": "UTF-8",
"star_events_count": 58,
"url": "https://raw.githubusercontent.com/longluo/leetcode/5a171f223c03cfdddb18488fd4bc5910039e21c4/Solutions/English/21. Merge Two Sorted Lists(Easy).md",
"visit_date": "2023-08-17T05:03:25.433075"
} | 3.390625 | stackv2 | # The Recursive Algorithm with Detailed Image Explanation
https://leetcode.com/problems/merge-two-sorted-lists/solutions/1857613/the-recursive-algorithm-with-intuition-image-explanation/
[Leetcode](https://leetcode.com/) [21. Merge Two Sorted Lists](https://leetcode.com/problems/merge-two-sorted-lists/).
***By Frank Luo***
# Intuition
We can use **Recursion** to solve this problem.
The key points of **Recursion** are $2$.
1. How to **terminate** the recursion: **Returns** when either $\texttt{l1}$ or $\texttt{l2}$ is $\texttt{null}$.
2. What to do in the process: if $\texttt{l1.val} \le \texttt{l2.val}$, then point $\texttt{l1.next}$ to the smaller of $\texttt{l1}$ and $\texttt{l2}$.
If $\texttt{l1.val} \le \texttt{l2.val}$, we can choose the smaller node, such as $\texttt{l1}$. However, the linked list is not reached the end, we will continue to compare.
Now we are compare $\texttt{l1.next}$ and $\texttt{l2}$. The $\texttt{l1.next}$ and $\texttt{l2}$ are processed in the recursive functions of the next layer.
We process such process and finally get the result.
# Image Explanation
You can get a **intuition** from below images.








```java
class Solution {
public ListNode mergeTwoLists(ListNode l1, ListNode l2) {
if (l1 == null) {
return l2;
} else if (l2 == null) {
return l1;
} else if (l1.val < l2.val) {
l1.next = mergeTwoLists(l1.next, l2);
return l1;
} else {
l2.next = mergeTwoLists(l1, l2.next);
return l2;
}
}
}
```
## Analysis
- **Time Complexity**: $O(m + n)$.
- **Space Complexity**: $O(m + n)$.
--------------------------
All suggestions are welcome.
If you have any query or suggestion please comment below.
Please upvote👍 if you like💗 it. Thank you:-)
Explore More [Leetcode Solutions](https://leetcode.com/discuss/general-discussion/1868912/My-Leetcode-Solutions-All-In-One). 😉😃💗
|
2024-11-18T18:31:53.513990+00:00 | 2019-04-17T15:48:20 | 9247d059f1390d696ecd3b48ae6d98daf150cd07 | 3 | {
"blob_id": "9247d059f1390d696ecd3b48ae6d98daf150cd07",
"branch_name": "refs/heads/master",
"committer_date": "2019-04-17T15:48:20",
"content_id": "18f7f927e4608f084e6083db4909b5b2458a1afc",
"detected_licenses": [
"MIT"
],
"directory_id": "639636501cdd8f258eeb877264b95caa76df28d9",
"extension": "md",
"filename": "README.md",
"fork_events_count": 2,
"gha_created_at": "2017-03-12T01:47:25",
"gha_event_created_at": "2023-08-28T14:55:16",
"gha_language": "JavaScript",
"gha_license_id": "MIT",
"github_id": 84692379,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 1404,
"license": "MIT",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:71",
"repo_name": "grxy/react-ad-block-detect",
"revision_date": "2019-04-17T15:48:20",
"revision_id": "45be390ae824baf66448fc95bef26aa3266a610c",
"snapshot_id": "30d5f3dd1d36ee7cce23c980f481746acbdbb724",
"src_encoding": "UTF-8",
"star_events_count": 10,
"url": "https://raw.githubusercontent.com/grxy/react-ad-block-detect/45be390ae824baf66448fc95bef26aa3266a610c/README.md",
"visit_date": "2023-09-01T06:19:41.152953"
} | 2.984375 | stackv2 | # react-ad-block-detect
A React component that renders its children if an ad blocker is detected
[](https://www.npmjs.com/package/react-ad-block-detect)
[](https://travis-ci.org/grxy/react-ad-block-detect)
[](https://circleci.com/gh/grxy/react-ad-block-detect)
[](https://coveralls.io/github/grxy/react-ad-block-detect?branch=master)
[](https://david-dm.org/grxy/react-ad-block-detect)
[](https://david-dm.org/grxy/react-ad-block-detect?type=dev)
## Installation
npm install react-ad-block-detect
## Usage
import React, { Component } from 'react';
import AdBlockDetect from 'react-ad-block-detect';
class MyComponent extends Component {
render() {
return (
<AdBlockDetect>
<p>Show this if an ad blocker has been enabled.</p>
</AdBlockDetect>
);
}
}
|
2024-11-18T18:31:53.554008+00:00 | 2016-10-04T18:14:41 | eda0bebb47fbd44950fc6ea7b9fc6113f2652ecf | 3 | {
"blob_id": "eda0bebb47fbd44950fc6ea7b9fc6113f2652ecf",
"branch_name": "refs/heads/master",
"committer_date": "2016-10-04T18:14:41",
"content_id": "e55a7a38477170032857bd53991e64f74a5dc839",
"detected_licenses": [
"MIT"
],
"directory_id": "caaa6f071d6dbe7224999e544fbf9a757f4c1117",
"extension": "md",
"filename": "README.md",
"fork_events_count": 1,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": 32019935,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 976,
"license": "MIT",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:72",
"repo_name": "MarcelloLins/MongoTools",
"revision_date": "2016-10-04T18:14:41",
"revision_id": "ef979bfb2a628dc42f358f059c56aae2fbf77ad5",
"snapshot_id": "59304e8345ad3924ac8a4f60cf20eada8f0e15f6",
"src_encoding": "UTF-8",
"star_events_count": 3,
"url": "https://raw.githubusercontent.com/MarcelloLins/MongoTools/ef979bfb2a628dc42f358f059c56aae2fbf77ad5/README.md",
"visit_date": "2020-04-06T07:17:16.228555"
} | 3.015625 | stackv2 | # MongoTools
A simple set of tools written for administration of MongoDB servers. All the tools available via this project works via CLI and were tested on Mono aswell, so you can run them on your unix environment if you feel like
This set of tools takes advantage of Parallel processing to perform each operation in it's own thread. Since MongoDB uses, for some operations, a "Collection-Level" lock,
by using one thread per collection processing, I can take full advantage of both the database performance and the network speed.
Read the Wiki for the parameters and examples of each tool.
Tools Available
======================
**[Export]** : Exports data from your MongoDB collections either as "CSV" (with a custom delimiter) or as "JSON".
**[Copy]** : Migrates data (and indexes) from one database to another (or from a server to another).
[Export]:https://github.com/MarcelloLins/MongoTools/wiki/Export
[Copy]:https://github.com/MarcelloLins/MongoTools/wiki/Copy |
2024-11-18T18:31:53.600499+00:00 | 2018-02-13T15:58:52 | 4c82b8d7cdd3ff6998064ab30ec0db00bade0f30 | 3 | {
"blob_id": "4c82b8d7cdd3ff6998064ab30ec0db00bade0f30",
"branch_name": "refs/heads/master",
"committer_date": "2018-02-13T15:58:52",
"content_id": "a5753ecf30517d6d84e515d396ba5c884f8ea35f",
"detected_licenses": [
"Apache-2.0"
],
"directory_id": "8cd89860f680cb6e0d21512844deaadaab17ce29",
"extension": "md",
"filename": "tips_for_scrum.md",
"fork_events_count": 0,
"gha_created_at": "2018-02-12T10:48:40",
"gha_event_created_at": "2018-02-13T15:56:49",
"gha_language": "HTML",
"gha_license_id": "Apache-2.0",
"github_id": 121234846,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 1212,
"license": "Apache-2.0",
"license_type": "permissive",
"path": "/tips_for_scrum.md",
"provenance": "stack-edu-markdown-0000.json.gz:73",
"repo_name": "scrum101/scrum101.github.io",
"revision_date": "2018-02-13T15:58:52",
"revision_id": "123a9c98545752c5e5d98623afb585ce44e22aa7",
"snapshot_id": "069b944582dc47ef153ec8ecc082d46f9101c386",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/scrum101/scrum101.github.io/123a9c98545752c5e5d98623afb585ce44e22aa7/tips_for_scrum.md",
"visit_date": "2021-04-30T15:13:53.734009"
} | 2.65625 | stackv2 | # Tips for running your scrum process and motivating your team
This page contains a list of tips which cover both team motivation and running the scrum process
## Team motivation
[Vision vs mission statement - what's the difference? and when to use them](visionvmission.md)
<div id="disqus_thread"></div>
<script>
/**
* RECOMMENDED CONFIGURATION VARIABLES: EDIT AND UNCOMMENT THE SECTION BELOW TO INSERT DYNAMIC VALUES FROM YOUR PLATFORM OR CMS.
* LEARN WHY DEFINING THESE VARIABLES IS IMPORTANT: https://disqus.com/admin/universalcode/#configuration-variables*/
/*
var disqus_config = function () {
this.page.url = PAGE_URL; // Replace PAGE_URL with your page's canonical URL variable
this.page.identifier = PAGE_IDENTIFIER; // Replace PAGE_IDENTIFIER with your page's unique identifier variable
};
*/
(function() { // DON'T EDIT BELOW THIS LINE
var d = document, s = d.createElement('script');
s.src = 'https://scrum101.disqus.com/embed.js';
s.setAttribute('data-timestamp', +new Date());
(d.head || d.body).appendChild(s);
})();
</script>
<noscript>Please enable JavaScript to view the <a href="https://disqus.com/?ref_noscript">comments powered by Disqus.</a></noscript>
|
2024-11-18T18:31:53.640532+00:00 | 2023-08-18T08:20:30 | b177c2bebd61309f807f0becd04efacc7c9c8c6e | 3 | {
"blob_id": "b177c2bebd61309f807f0becd04efacc7c9c8c6e",
"branch_name": "refs/heads/main",
"committer_date": "2023-08-18T08:20:30",
"content_id": "66d91cc38ae1c6d94c9314ab6cee67bb5bcc322c",
"detected_licenses": [
"MIT"
],
"directory_id": "dbe739d138d530a48b0d8bb1539303c96a3c23c3",
"extension": "md",
"filename": "orderlineamountdiscountrewardprovidersettingsbase.md",
"fork_events_count": 885,
"gha_created_at": "2012-06-15T09:18:34",
"gha_event_created_at": "2023-09-14T14:31:37",
"gha_language": null,
"gha_license_id": "MIT",
"github_id": 4673374,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 1077,
"license": "MIT",
"license_type": "permissive",
"path": "/12/umbraco-commerce/reference/umbraco-commerce-core/umbraco-commerce-core-discounts-rewards/orderlineamountdiscountrewardprovidersettingsbase.md",
"provenance": "stack-edu-markdown-0000.json.gz:74",
"repo_name": "umbraco/UmbracoDocs",
"revision_date": "2023-08-18T08:20:30",
"revision_id": "9142cdb9ac2594b23fee01c1c438604b215215af",
"snapshot_id": "c4c4995b0fca6beb63ebf9c375edb1b8a10af165",
"src_encoding": "UTF-8",
"star_events_count": 232,
"url": "https://raw.githubusercontent.com/umbraco/UmbracoDocs/9142cdb9ac2594b23fee01c1c438604b215215af/12/umbraco-commerce/reference/umbraco-commerce-core/umbraco-commerce-core-discounts-rewards/orderlineamountdiscountrewardprovidersettingsbase.md",
"visit_date": "2023-08-18T12:42:54.647750"
} | 2.859375 | stackv2 | ---
title: OrderLineAmountDiscountRewardProviderSettingsBase
description: API reference for OrderLineAmountDiscountRewardProviderSettingsBase in Umbraco Commerce
---
## OrderLineAmountDiscountRewardProviderSettingsBase
```csharp
public abstract class OrderLineAmountDiscountRewardProviderSettingsBase
```
**Namespace**
* [Umbraco.Commerce.Core.Discounts.Rewards](README.md)
### Constructors
#### OrderLineAmountDiscountRewardProviderSettingsBase
The default constructor.
```csharp
public OrderLineAmountDiscountRewardProviderSettingsBase()
```
### Properties
#### AdjustmentType
```csharp
public AdjustmentType AdjustmentType { get; set; }
```
---
#### Amounts
```csharp
public IDictionary<Guid, decimal?> Amounts { get; set; }
```
---
#### AmountsIncludeTax
```csharp
public bool AmountsIncludeTax { get; set; }
```
---
#### Percentage
```csharp
public decimal? Percentage { get; set; }
```
---
#### PriceType
```csharp
public OrderLinePriceType PriceType { get; set; }
```
<!-- DO NOT EDIT: generated by xmldocmd for Umbraco.Commerce.Core.dll -->
|
2024-11-18T18:31:53.676343+00:00 | 2023-08-11T17:58:17 | 510e0ac864f94fab564128c22f36bae3f819f0fb | 3 | {
"blob_id": "510e0ac864f94fab564128c22f36bae3f819f0fb",
"branch_name": "refs/heads/master",
"committer_date": "2023-08-11T17:58:17",
"content_id": "bda3c0546603000d185b32de94c5c49b95c0ccc5",
"detected_licenses": [
"MIT"
],
"directory_id": "54476db0f6edccd05ff8364caecb66ce04cbd4c3",
"extension": "md",
"filename": "2019-08-19-aprameyo-pal.md",
"fork_events_count": 2,
"gha_created_at": "2017-01-31T05:24:40",
"gha_event_created_at": "2023-04-12T09:12:53",
"gha_language": "HTML",
"gha_license_id": "MIT",
"github_id": 80494033,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 866,
"license": "MIT",
"license_type": "permissive",
"path": "/_seminars/2019/2019-08-19-aprameyo-pal.md",
"provenance": "stack-edu-markdown-0000.json.gz:75",
"repo_name": "siddhartha-gadgil/DeptWeb",
"revision_date": "2023-08-11T17:58:17",
"revision_id": "0df0ae6e67217487ab87686cdd47b798fe9cc4a5",
"snapshot_id": "f42863924e86231477dbb89822b0c687e722d462",
"src_encoding": "UTF-8",
"star_events_count": 2,
"url": "https://raw.githubusercontent.com/siddhartha-gadgil/DeptWeb/0df0ae6e67217487ab87686cdd47b798fe9cc4a5/_seminars/2019/2019-08-19-aprameyo-pal.md",
"visit_date": "2023-08-16T04:19:56.816172"
} | 2.53125 | stackv2 | ---
speaker: Aprameyo Pal (University of Duisburg-Essen, Germany)
title: "A central value formula of degree 6 complex L-series and arithmetic applications"
date: 19 August, 2019
time: 3 pm
venue: LH-1, Mathematics Department
---
We prove an explicit central value formula for a family of complex L-series
of degree 6 for GL2 × GL3 which arise as factors of certain Garret--Rankin
triple product L-series associated with modular forms. Our result generalizes
a previous formula of Ichino involving Saito--Kurokawa lifts, and as an
application, we prove Deligne's conjecture about the algebraicity of the
central values of the considered L-series up to the relevant periods. I would
also include some other arithmetic applications towards the subconvexity problem,
construction of associated p-adic L function, etc.
This is joint work with Carlos de Vera Piquero.
|
2024-11-18T18:31:53.718169+00:00 | 2020-06-15T18:45:27 | be09a7b06c8300d46464377706b0564113fcb161 | 3 | {
"blob_id": "be09a7b06c8300d46464377706b0564113fcb161",
"branch_name": "refs/heads/master",
"committer_date": "2020-06-15T18:45:27",
"content_id": "41775470ad02679f7a0cb454184e8c62432fc858",
"detected_licenses": [
"MIT"
],
"directory_id": "13f01282a59fcfa24eef8652715e790a4d08d735",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": 271608567,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 2364,
"license": "MIT",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:76",
"repo_name": "cryocaldera/abuskosi",
"revision_date": "2020-06-15T18:45:27",
"revision_id": "84bb45c60851ac6cfdd26e80ed710db15e9786f5",
"snapshot_id": "0acd750be93fad1382a1cb0f4388014543447e8e",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/cryocaldera/abuskosi/84bb45c60851ac6cfdd26e80ed710db15e9786f5/README.md",
"visit_date": "2022-11-07T04:46:54.408292"
} | 3.015625 | stackv2 | # abuskosi
ABUSKOSI shall become a universal programming environment for strategic (wargames and) conflict simulations.
The name itself actually is an acronym of this. It's just been translated to another language: "AusgangsBasis für Universelle Strategische (kriegsspiele und) KOnfliktSImulationen".
### Just a remark
Right now a lot of docs are written down in german only (especially project planning stuff), but don't feel scared about this.
It's just that currently all contributors are natively speaking german, so we simply decided to ease our lives. :)
Nevertheless, on the long run we want to address international audiance and contributors.
Our source code comments are already english. (Feel free to blame me if you find any other and I will correct this!)
The reminder of the project shall follow to "international-mode" when either:
* we feel our project has a certain degree of maturity or
* interesting contributions start to drop in from the international area or
* there is any reasonable interest outside of DE/AT/CH on this project
## Project goals
On the long run the targets of this project are:
* to provide a universal framework for strategic wargames and conflict simulations
(i.a. shall it serve as a unified foundation for various games the project initiator had in mind like "NeroRost"-universes.)
* establish and maintain a cross platform compability, meaning both developers (Linux, Windows, macOS) and gamers (Linux, Windows, macOS, later including: Retropie, Android, maybe Playstation, etc.) alike
* level up the knowhow and technical experise of all (core-)developers
Further project goals that depend on the outcome of several evaluation phases:
* Either commit to an already existing (open source) game engine
* Or continue its development as a fork
* Or develop an entirely new game engine
## Versioning
We use [SemVer](http://semver.org/) for versioning. For the versions available, see the [tags on this repository](https://github.com/cryocaldera/abuskosi/tags).
## Contributing
Pull requests are welcome. For major changes, please open an issue first to discuss what you would like to change.
Please make sure to:
* update tests as appropriate
* respect existing naming conventions, file and folder structures, etc. (see docs folders for details)
## License
[MIT](https://choosealicense.com/licenses/mit/)
|
2024-11-18T18:31:53.930461+00:00 | 2015-09-18T15:22:37 | aba72f63f58f96643c2354e91aae2f7fff9ad72b | 4 | {
"blob_id": "aba72f63f58f96643c2354e91aae2f7fff9ad72b",
"branch_name": "refs/heads/master",
"committer_date": "2015-09-18T15:22:37",
"content_id": "594b4c63f19f857106d5810551009cb4b11718b7",
"detected_licenses": [
"MIT"
],
"directory_id": "0738995248f098add821d53cf4460fb8e71c5ed3",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": null,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 1785,
"license": "MIT",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:77",
"repo_name": "iDay/react-native-lightbox",
"revision_date": "2015-09-18T15:22:37",
"revision_id": "377f3b72d94335d5dc999532a2ac3e02c4f603c4",
"snapshot_id": "2ddf8590c480534482753742c8e9a8ee2dae62c0",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/iDay/react-native-lightbox/377f3b72d94335d5dc999532a2ac3e02c4f603c4/README.md",
"visit_date": "2021-01-15T20:33:14.600598"
} | 3.75 | stackv2 | # react-native-lightbox
## Installation
```
npm install --save react-native-lightbox
```
**This module requires React Native 0.11 or later**
## Usage
```js
var Lightbox = require('react-native-lightbox');
var LightboxView = React.createClass({
render: function() {
return (
<Lightbox>
<Image
style={{ height: 300 }}
source={{ uri: 'http://knittingisawesome.com/wp-content/uploads/2012/12/cat-wearing-a-reindeer-hat1.jpg' }}
/>
</Lightbox>
);
},
});
```
### Manual Opening & Closing
The component exposes the `open` and `close` methods. Smack a `ref` on the `<Lightbox>` and you're good to go. This is probably quite useful if you're doing a custom header.
```js
render: function() {
return (
<Lightbox ref="lightbox">…</Lightbox>
);
},
handleSomething: function() {
this.refs.lightbox.open();
}
```
## Properties
| Prop | Type | Description |
|---|---|---|
|**`activeProps`**|`object`|Optional set of props applied to the content component when in lightbox mode. Usable for applying custom styles or higher resolution image source.|
|**`header`**|`element`|Custom header instead of default with X button|
|**`onClose`**|`function`|Triggered when lightbox is closed|
|**`onOpen`**|`function`|Triggered when lightbox is opened|
|**`underlayColor`**|`string`|Color of touchable background, defaults to `black`|
|**`swipeToDismiss`**|`bool`|Enables gestures to dismiss the fullscreen mode by swiping up or down, defaults to `true`.|
## Demo

## Example
Check full example in the `Example` folder.
## License
[MIT License](http://opensource.org/licenses/mit-license.html). © Joel Arvidsson
|
2024-11-18T18:31:54.026824+00:00 | 2021-01-26T09:13:36 | bfeafe12f5f33337906c88a04d602dfc98c79bf0 | 3 | {
"blob_id": "bfeafe12f5f33337906c88a04d602dfc98c79bf0",
"branch_name": "refs/heads/master",
"committer_date": "2021-01-26T09:13:36",
"content_id": "f115fe25aaf793598680b532e6556cc6361a8b03",
"detected_licenses": [
"MIT"
],
"directory_id": "4bfedfdf5ce5429b0cd14f7e2fd5a1e700e72cb5",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": "2020-03-25T17:19:19",
"gha_event_created_at": "2020-03-25T17:19:20",
"gha_language": null,
"gha_license_id": "MIT",
"github_id": 250043719,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 2323,
"license": "MIT",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:78",
"repo_name": "lusk/vue-i18n-extract",
"revision_date": "2021-01-26T09:13:36",
"revision_id": "3766ae8420688c60e032c4a34be3423f216720ef",
"snapshot_id": "c56acdd66120555c00daa07d7118121135d064e8",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/lusk/vue-i18n-extract/3766ae8420688c60e032c4a34be3423f216720ef/README.md",
"visit_date": "2021-04-24T00:26:42.362627"
} | 3.140625 | stackv2 | <h1 align="center">vue-i18n-extract</h1>
<p align="center">
<img align="center" src="https://raw.githubusercontent.com/pixari/vue-i18n-extract/master/demo/screenshots/vue-i18n-extract-3.png">
</p>
<p align="center">
<a href="https://www.npmjs.com/package/vue-i18n-extract"><img src="https://img.shields.io/npm/v/vue-i18n-extract.svg?style=flat-square" alt="NPM Version"></a>
<a href="https://www.npmjs.com/package/vue-i18n-extract"><img src="https://img.shields.io/npm/dm/vue-i18n-extract.svg?style=flat-square" alt="Downloads"></a>
<a href="https://circleci.com/gh/pixari/vue-i18n-extract"><img src="https://circleci.com/gh/pixari/vue-i18n-extract/tree/master.png?style=shield" alt="CircleCI Status"></a>
<a href="https://snyk.io/test/github/pixari/vue-i18n-extract?targetFile=package.json"><img src="https://snyk.io/test/github/pixari/vue-i18n-extract/badge.svg?targetFile=package.json" alt="Known Vulnerabilities"></a>
<a href="https://codeclimate.com/github/pixari/vue-i18n-extract/maintainability"><img src="https://api.codeclimate.com/v1/badges/d21f341c33b2bfb6fe0e/maintainability" alt="Maintainability"></a>
</p>
---
`vue-i18n-extract` is built to work with your Vue.js projects using [vue-i18n](https://kazupon.github.io/vue-i18n/). When run `vue-18n-extract` analyses your Vue.js source code for any `vue-i18n` key usages (ex. $t(''), $tc(''), ...) as well as your language files (ex. de_DE.js, en_EN.json, ...), in order to:
- [x] I18NReport keys that are missing in the language files.
- [x] I18NReport unused keys in the language files.
## :book: Documentation
Check the official documentation here: https://pixari.github.io/vue-i18n-extract/
## :exclamation: Issues
I'm sure you'll find bugs and when you do it would be great if you'd could [report them here](https://github.com/pixari/vue-i18n-extract/issues).
## :muscle: Contribution
The project is still in its early stages and in progress. I think there's no need for guidelines yet, so feel free to contribute or give feedback as you prefer.
PLEASE: don't forget to update the documentation if you change the API. (https://github.com/pixari/vue-i18n-extract/tree/master/docs)
## :copyright: License
[MIT](http://opensource.org/licenses/MIT)
|
2024-11-18T18:31:54.092398+00:00 | 2018-07-11T14:13:44 | a844540dc6398bf542fa2271d56c6690339b17d7 | 2 | {
"blob_id": "a844540dc6398bf542fa2271d56c6690339b17d7",
"branch_name": "refs/heads/master",
"committer_date": "2018-07-11T14:13:44",
"content_id": "14181258380ab0621bfd76eaf683bb10b33ddb58",
"detected_licenses": [
"MIT"
],
"directory_id": "24447615e9256854f348b4b642df2a446d9d4c8e",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": "2018-07-11T14:12:29",
"gha_event_created_at": "2018-07-11T14:12:30",
"gha_language": null,
"gha_license_id": null,
"github_id": 140584854,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 554,
"license": "MIT",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:79",
"repo_name": "vime/SBEMimage",
"revision_date": "2018-07-11T14:13:44",
"revision_id": "0a59b39619303fc1bbe256acecfd901593c9ca86",
"snapshot_id": "0f34b3201785b621d5e261e7ee86f3ddf7e8f03e",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/vime/SBEMimage/0a59b39619303fc1bbe256acecfd901593c9ca86/README.md",
"visit_date": "2020-03-22T20:14:53.714628"
} | 2.5 | stackv2 | # SBEMimage
Acquisition software for serial block-face electron microscopy, made with Python and PyQt
## Getting started
Please read SBEMimage_Manual.pdf (in the project root folder). It currently contains installation instructions and a short introduction to the software. The document will be expanded into a full manual soon.
## Contributing
## Authors
* **Benjamin Titze** - *Design and programming* - [btitze](https://github.com/btitze)
## License
This project is licensed under the MIT License - see the [LICENSE](LICENSE) file for details
|
2024-11-18T18:31:54.421046+00:00 | 2019-02-12T21:57:47 | 7cd8df7e8f40a926847d153ed589a0926e47bd72 | 4 | {
"blob_id": "7cd8df7e8f40a926847d153ed589a0926e47bd72",
"branch_name": "refs/heads/master",
"committer_date": "2019-02-12T21:57:47",
"content_id": "bdf08945ba68a54a52e7d0e32842a8714bcbce0b",
"detected_licenses": [
"MIT"
],
"directory_id": "784e7cade958d7ef10c346f88a10ef9851fa25ec",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": 170396981,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 1976,
"license": "MIT",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:80",
"repo_name": "alexliubj/Smoke-Framework-Example",
"revision_date": "2019-02-12T21:57:47",
"revision_id": "e2172b42f3ecd47ab164a329b725eb2665153212",
"snapshot_id": "fd33b6b89f537e8513fd11da913ae5e4305777bf",
"src_encoding": "UTF-8",
"star_events_count": 2,
"url": "https://raw.githubusercontent.com/alexliubj/Smoke-Framework-Example/e2172b42f3ecd47ab164a329b725eb2665153212/README.md",
"visit_date": "2020-04-22T13:05:48.529605"
} | 3.859375 | stackv2 | ## How to make a simple Post request with Smoke-Framework
### Start
First of all, open terminal and create a new empty folder.
Create an executable package by running the following commands in the terminal:
`swift package init --type executable`
Open `Package.swift` file and add `smoke-framework` into dependencies section and also update targets dependencies:
```swift
let package = Package(
name: "SmokeAPI",
dependencies: [
.package(url: "https://github.com/amzn/smoke-framework.git", .upToNextMajor(from: "0.6.0"))
],
targets: [
// Targets are the basic building blocks of a package. A target can define a module or a test suite.
// Targets can depend on other targets in this package, and on products in packages which this package depends on.
.target(
name: "SmokeAPI",
dependencies: ["SmokeOperations", "SmokeHTTP1"]),
.testTarget(
name: "SmokeAPITests",
dependencies: ["SmokeAPI"]),
]
)
```
### Build
Back to terminal and type `swift build`. This command will fetch `Smoke-framework` for you into this folder as well as those dependencies required by `Smoke-framework`.
You will see `main.swift` file locates in `Sources/SmokeAPI` and if you open it there will be only one line of code: `print("Hello, world!")`
In this example, there are a few files which are abstracted from test cases in `Smoke-framework`.
* `main` sets up the `post` API and starts service
* `ExampleInput` defines input data structure
* `ExampleError` defines errors
* `ExampleOutput` defines output structure
### Run
Type `swift run SmokeAPI` in terminal to run main file and start service.
### Test
You can either test it in your own project with HTTP request or test it with some tools like postman.
Here's the screenshot of testing it with postman

|
2024-11-18T17:55:35.004609+00:00 | 2018-06-06T15:49:12 | f9214d34a36c82c2c248ed9cf08db2ba1acd38e3 | 4 | {
"blob_id": "f9214d34a36c82c2c248ed9cf08db2ba1acd38e3",
"branch_name": "refs/heads/master",
"committer_date": "2018-06-06T15:49:12",
"content_id": "a881d2bb0dcae6fd878e3fe914a9f2fb2f13ad85",
"detected_licenses": [
"Apache-2.0"
],
"directory_id": "00b1c5585d66a231f4cf9fe0ec1160d99e6e1de1",
"extension": "md",
"filename": "osgiaas-cli-clojure.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": null,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 1414,
"license": "Apache-2.0",
"license_type": "permissive",
"path": "/docs/cli/osgiaas-cli-clojure.md",
"provenance": "stack-edu-markdown-0000.json.gz:81",
"repo_name": "avatar21/osgiaas",
"revision_date": "2018-06-06T15:49:12",
"revision_id": "f0e3d84f91bde64809e5fbe89ed30c6d9a43b0ab",
"snapshot_id": "27cbf504bcd87e1d88b1b1dbd138bc18f1cd258c",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/avatar21/osgiaas/f0e3d84f91bde64809e5fbe89ed30c6d9a43b0ab/docs/cli/osgiaas-cli-clojure.md",
"visit_date": "2020-08-09T08:28:31.963006"
} | 3.859375 | stackv2 | # OSGiaaS CLI Clojure Module
A module that implements a Clojure REPL.
## Usage
Simply type `clj` followed by any Clojure code.
For example:
```
>> clj (defn double [x] (* x 2))
< #'clojure.core/double
>> clj (double 4)
< 8
```
When run through pipes, the Clojure code should return a function that takes each input line as an argument,
returning something to be printed (nothing is printed when the function returns nil).
For example, to only print the lines containing the word 'text' from the output of some_command:
```
>> some_command | clj (fn [line] (if (.contains line "text") line nil))
```
The following variables are always available in the "user" namespace:
* out - the command output stream.
* err - the command error stream.
* ctx - the Clojure command's OSGi service ComponentContext.
For example, to check the implementation class of the OSGi `ComponentContext`:
```
>> (.getClass user/ctx)
< class org.apache.felix.scr.impl.manager.ComponentContextImpl
```
To access REPL functions, you can require the REPL namespace like this:
```
>> clj (require '[clojure.repl :as repl])
```
You can then access docs, for example:
```
>> clj (repl/doc +)
```
## System requirements
The Clojure jar is expected to be available in the runtime.
With the osgi-run plugin, you can achieve that by adding this dependency to your project:
```groovy
systemLib 'org.clojure:clojure:1.8.0'
```
|
2024-11-18T17:55:35.124553+00:00 | 2018-02-18T08:25:56 | f89566d45ea382a5794a55dc3574844f7c6a4cd1 | 4 | {
"blob_id": "f89566d45ea382a5794a55dc3574844f7c6a4cd1",
"branch_name": "refs/heads/master",
"committer_date": "2018-02-18T08:25:56",
"content_id": "4e6936c50dfc13eec7afa8956a4bbf56cfabb20b",
"detected_licenses": [
"Unlicense"
],
"directory_id": "5d8de85c8fb328a30c0efd6c3fe441b90b6a0055",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": null,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 1899,
"license": "Unlicense",
"license_type": "permissive",
"path": "/packer/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:82",
"repo_name": "hostyStack/gitlab-runner-packer",
"revision_date": "2018-02-18T08:25:56",
"revision_id": "f7eeb1f2e210a0c15c180dddd5b9cc8b78488275",
"snapshot_id": "ed1379fb7fdab420b726b8f65ffb593f3c75eaa6",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/hostyStack/gitlab-runner-packer/f7eeb1f2e210a0c15c180dddd5b9cc8b78488275/packer/README.md",
"visit_date": "2021-09-07T05:48:23.534306"
} | 3.65625 | stackv2 | # Gitlab Runner AMI
Builds and configures an AMI with:
**Infra**
* Updates to latest packages and enables auto update of security packages
* firewall allowing only ssh incoming
**Tools**
* docker
* kubectl
* aws cli
* gitlab runner
* jq
* docker cleanup jobs
**Initialisation**
* Gitlab runner configuration - see configuration below
* Kubectl configuration setup - see configuration below
## Build the AMI
```bash
packer build -var 'aws_region=us-east-1' template.json
```
## Run configuration
The initialisation script `scripts/init/runner-init.sh` should be executed when the instance is first run. It requires the following environment variables set.
* **RUNNER_TAG_LIST:** A comma separated list of tags for the runner
* **RUNNER_NAME:** A name to give runner in gitlab
* **REGISTRATION_TOKEN:** The project registration token
* **KUBE_CONFIG_PATH** A path to a kubectl config file within a S3 bucket
The following is an example terraform resource that uses the aws `user_data` configuration value to configure and execute the copied `runner-init.sh` script.
```
resource "aws_instance" "gitlab-runner" {
count = "1"
ami = "${data.aws_ami.gitlab-runner.id}"
instance_type = "t2.micro"
security_groups = ["${aws_security_group.gitlab-runner.name}"]
user_data = <<EOF
#!/bin/bash
export RUNNER_TAG_LIST=YOUR_PROJECT_NAME,shell,docker,kubectl,aws
export RUNNER_NAME=myproject-gitlab-runner
export REGISTRATION_TOKEN=YOUR_PROJECT_TOKEN
export KUBE_CONFIG_PATH=s3://YOUR_BUCKET/kubectl/config
/opt/runner-init.sh
EOF
iam_instance_profile="${aws_iam_instance_profile.gitlab-runner.name}"
key_name = "${aws_key_pair.gitlab-runner.key_name}"
associate_public_ip_address = "true"
tags {
Name = "gitlab-runner"
app = "gitlab"
Environment = "engineering"
}
}
```
Also see the sample terraform project included in this repo. |
2024-11-18T17:55:35.192188+00:00 | 2022-04-19T15:38:21 | c1c88b6225d1ac57ac29e010237b55e86d5746ba | 3 | {
"blob_id": "c1c88b6225d1ac57ac29e010237b55e86d5746ba",
"branch_name": "refs/heads/master",
"committer_date": "2022-04-19T15:38:21",
"content_id": "212f5d865ec37e517c857036be348f8bec3f99c6",
"detected_licenses": [
"MIT"
],
"directory_id": "b34402fe1f15bd0876395296d6bb9fd6cc599e59",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": "2019-10-26T09:25:58",
"gha_event_created_at": "2020-04-08T14:30:00",
"gha_language": "Java",
"gha_license_id": "MIT",
"github_id": 217685773,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 7473,
"license": "MIT",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:83",
"repo_name": "lqbilbo/Princeton-Algorithm-from-Robert-Sedgewick",
"revision_date": "2022-04-19T15:38:21",
"revision_id": "3d52942189923d1e09924145d056e089ff5d39d3",
"snapshot_id": "f18fb6bbb8c1d4739d9313f5f96e3afb1015b17f",
"src_encoding": "UTF-8",
"star_events_count": 1,
"url": "https://raw.githubusercontent.com/lqbilbo/Princeton-Algorithm-from-Robert-Sedgewick/3d52942189923d1e09924145d056e089ff5d39d3/README.md",
"visit_date": "2022-05-01T16:15:35.970107"
} | 3.203125 | stackv2 | # Princeton-Algorithm-from-Robert-Sedgewick
learning algorithm from Robert Sedgewick
# Lesson 1st Union-Find
1. dynamic connectivity
2. quick find
3. quick union
4. improvements
5. applications
## Subtext of the lecture
- Steps to developing a usable algorithm.
- Model the problem.
- Find an algorithm to solve it.
- Fast enough? Fits in memory?
- If not, figure out why.
- Find a way to address the problem.
- Iterate until satisfied.
- The scientific method.
- Mathematical analysis.
# Lesson 2nd Bags, Queues, and Stacks
1. stacks
2. resizing arrays
3. queues
4. generics
5. iterators
6. applications
- stack
- void push(String item)
- String pop()
- boolean isEmpty()
## stack implementations: resizing array vs. linked list
- Tradeoffs. Can implement a stack with either resizing array or linked list;
client can use interchangeably. Which one is better?
Linked-list implementation.
- Every operation takes constant time in the worst case.
- Uses extra time and space to deal with the links.
Resizing-array implementation.
- Every operation takes constant amortized time.
- Less wasted space.
- queue
- void enqueue(String item)
- String dequeue()
- boolean isEmpty()
- bag
- void add(Item x)
- int size()
- Iterable<Item> iterator()
# Lesson 3rd Algorithm Analysis
1. mathematical models
2. order-of-growth classifications
3. theory of algorithms
4. memory
the mostly common running time:
1, log N, N, NlogN, N^2, N^3, and 2^N
binary Search implement in Java - Arrays.binarySearch
## Theory of algorithms
1. Goals.
- Establish "difficulty" of a problem.
- Develop "optimal" algorithms
2. Approach.
- Suppress details in analysis: analyze "to within a constant factor".
- Eliminate variability in input model by focusing on the worst case.
3. Optimal algorithm.
- Performance guarantee (to within a constant factor) for any input.
- No algorithm can provide a better performance guarantee.
example: Three-sum problem
## Algorithm design approach
Start.
- Develop an algorithm
- Prove a lower bound.
Gap?
- Lower the upper bound (discover a new algorithm).
- Raise the lower bound (more difficult).
Golden Age of Algorithm Design.
- 1970s-.
- Steadily decreasing upper bounds for many important problems.
- Many known optimal algorithms.
Caveats.
- Overly pessimistic to focus on worst case?
- Need better than "to within a constant factor" to predict performance.
## Memory
typical memory usage for objects in Java
1. Object overhead: 16bytes
2. Reference: 8bytes
3. Padding: Each object uses a multiple of 8 bytes.
- Primitive type: 4 bytes for int, 8 bytes for double, ...
- Object reference: 8 bytes.
- Array: 24 bytes + memory for each array entry.
- Object: 16 bytes + memory for each instance variable + 8 if
- inner class (for pointer to enclosing class).
- Padding: round up to multiple of 8 bytes.
Shallow memory usage: Don't count referenced objects.
Deep memory usage: If array entry or instance variable is a reference,
add memory (recursively) for referenced object.
Turning the crank: summary
________________________________
1.Empirical analysis.
2.Mathematical analysis.
3.Scientific method.
# Lesson 4th Elementary Sorts
## selection sort
- In iteration i, find index min of smallest remaining entry.
- Swap a[i] and a[min]
## insertion sort
- In iteration i, swap a[i] with each larger entry to its left.
## shuffling
- Generate a random real number for each array entry.
- Sort the array.
## convex hull
The convet hull of a set of N points is the smallest perimeter fence enclosing the points.
- mechanical algorithm.
Hammer nails perpendicular to plane; stretch elastic rubber band around points.
- application: motion planning.
Robot motion planning. Find shortest path in the plane from s to t that avoids a polygonal obstacle.
- application: farthest pair.
Civen N points in the plane, find a pair of points with the largest Euclidean distance between them.
# Lesson 5th mergesort
## mergesort
Basic plan.
- Divide array into two havles.
- Recursively sort each half.
- Merge two halves.
## bottom-up mergesort
Basic plan.
- Pass through array, merging subarrays of size 1.
- Repeat for subarrays of size 2, 4, 8, 16..
## sorting complexity
Computational complexity. Framework to study efficientcy of algorithms
for solving a particular problem X.
Lower bound may not hold if the algorithm has information about:
- The initial order of the input.
- The distribution of key values.
- The representation of the keys.
## stability
- insertion sort is stable.
- selection sort is not stable.
- shell sort is not stable.
- merge sort is stable.
# Lesson 6th quicksort
## quicksort
Basic plan.
- Shuffle the array. Shuffling is needed for performance guarantee.
- Partition so that, for some j
- entry a[j] is in place
- no larger entry to the left of j
- no smaller entry to the right of j
- Sort each piece recursively.
### performance characteristics
Worst case. Number of compares is quadratic
Average case. Number of compares is ~ 1.39 NlgN.
- 39% more compares than mergesort.
- But faster than mergesort in practice because of less data movement.
Random shuffle
- Probabilistic guarantee against worst case.
- Basis for math model that can be validated with experiments.
### Quicksort properties
Proposition. Quicksort is an in-place sorting algorithm.
- Partitioning: constant extra space.
- Depth of recursion: logarithmic extra space (with high probability)
Proposition. Quicksort is not ~stable~.
Median of sample.
- Best choice of pivot them = median
- Estimate true median by taking median of sample.
- Median-of-3(random) items.
### Quicksort selection
Goal. Given an array of N items, find the top k largest
Ex. Min(k=0), max(k=N-1), median(k=N/2)
Applications.
- Order statistics.
- Find the 'top k'.
Repeat in one subarray, depending on j; finished when j equals k.
Proposition. Compare-based selection algorithm whose worst-case running time is linear.
### Quicksort duplicate keys
Mistake. Put all items equal to the partitioning item on one side.
Consequence. ~ 1/2 N^2 compares when all keys equal.
Recommended. Stop scans on itmes equal to the partitioning item.
Consequence. ~NlgN compares when all keys equal.
### 3-way partitioning
- let v be partiitnoning item a[lo]
- Scan i from left to right.
- (a[i] < v): exchange a[lt] with a[i]: increment both lt and i
- (A[i] > v): exchange a[gt] with a[i]: decrement gt
- (a[i] == v): increment i
### system sorts
- Java
Arrays.sort()
- C
qsort()
### which algorithm to use?
Internal sorts:
- Insertion sort, selection sort, bubble sort, shaker sort.
- Quick sort, merge sort, heap sort, sample sort, shell sort.
- Solitaire sort, red-black sort, splay sort, Yaroslavskiy sort, p-sort,...
External sorts. Poly-phase merge sort, cascade-merge, oscillating sort.
# Lesson 7th Priority queues
## API and elementary implementations
Collections.. Insert and delete items. Which item to delete?
Stack. Remove the item most recently added.
Queue. Remove the item least recently added.
Randomized queue. Remove a random item.
Priority queue. Remove the largest(or smallest) item.
### API.
Requirement。 Generic items are Comparable.
```
MaxPQ()
void insert(Key v)
Key delMax()
boolean isEmpty()
```

|
2024-11-18T17:58:59.543957+00:00 | 2023-07-14T05:41:58 | b9e434c77cbc6e025c9fb48d090c16f8e6efbb2a | 3 | {
"blob_id": "b9e434c77cbc6e025c9fb48d090c16f8e6efbb2a",
"branch_name": "refs/heads/main",
"committer_date": "2023-07-14T05:41:58",
"content_id": "b2671ae1644500fa80c74f6469d6c105ce8ae6c2",
"detected_licenses": [
"MIT"
],
"directory_id": "c0dea6c80b243cfcb0e20caeaad15f34ad573b8c",
"extension": "md",
"filename": "README.md",
"fork_events_count": 417,
"gha_created_at": "2020-09-01T18:19:20",
"gha_event_created_at": "2023-07-18T23:28:00",
"gha_language": "TypeScript",
"gha_license_id": "MIT",
"github_id": 292074617,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 10119,
"license": "MIT",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:84",
"repo_name": "vn7n24fzkq/github-profile-summary-cards",
"revision_date": "2023-07-14T05:41:58",
"revision_id": "7e359c68504a50105c54073afe41b2c4385d6274",
"snapshot_id": "04937227b30d46924b8829751d8b70802d4bc4c4",
"src_encoding": "UTF-8",
"star_events_count": 2164,
"url": "https://raw.githubusercontent.com/vn7n24fzkq/github-profile-summary-cards/7e359c68504a50105c54073afe41b2c4385d6274/README.md",
"visit_date": "2023-07-24T07:05:01.276596"
} | 2.84375 | stackv2 | <div align="center">
<h1>GitHub Profile Summary Cards</h1>
[繁體中文](./docs/README.zh-tw.md)
<p>
A tool to generate your github summary card for profile README. Inspired by <a href=https://github.com/tipsy/profile-summary-for-github>profile-summary-for-github</a>
</p>
<p>
:star: This repo is just for fun, feel free to contribute! :star:
</p>
<p align="center">
<a href="https://github.com/vn7n24fzkq/github-profile-summary-cards/stargazers">
<img alt="Stargazers" src="https://img.shields.io/github/stars/vn7n24fzkq/github-profile-summary-cards?style=for-the-badge&logo=github&color=f4dbd6&logoColor=D9E0EE&labelColor=302D41"></a>
<a href="https://github.com/vn7n24fzkq/github-profile-summary-cards/releases/latest">
<img alt="Releases" src="https://img.shields.io/github/release/vn7n24fzkq/github-profile-summary-cards.svg?style=for-the-badge&logo=semantic-release&color=f5bde6&logoColor=D9E0EE&labelColor=302D41"/></a>
<a href="https://www.conventionalcommits.org/en/v1.0.0/">
<img alt="conventionalcommits" src="https://img.shields.io/badge/Conventional%20Commits-1.0.0-%23FE5196?style=for-the-badge&logo=conventionalcommits&color=ee99a0&logoColor=D9E0EE&labelColor=302D41"></a>
<a href="https://github.com/vn7n24fzkq/github-profile-summary-cards/actions/workflows/github-action.yml">
<img alt="testandlint" src="https://img.shields.io/github/actions/workflow/status/vn7n24fzkq/github-profile-summary-cards/test-and-lint.yml?branch=main&label=Test%20and%20Lint&style=for-the-badge&color=a6da95"></a>
</p>
</div>
<div align="center">
<p>
<a href="https://github-profile-summary-cards.vercel.app/demo.html">Get your own cards now!!</a>
</p>





</div>
## Themes
| | | | | |
|:---:|:---:|:---:|:---:|:---:|
|default|2077|dracula|github|github_dark|
||| |||
|gruvbox|monokai|nord_bright|nord_dark|radical|
||| | ||
|solarized|solarized_dark|tokyonight|vue|zenburn|
||| | ||
|transparent|
||
[More themes](https://github.com/vn7n24fzkq/github-profile-summary-cards-example/tree/master/profile-summary-card-output)
## How to use (API)
### Profile details card

`http://github-profile-summary-cards.vercel.app/api/cards/profile-details?username={username}&theme={theme_name}`
- Accept url parameters
- theme
- Theme name
- username
- Username
### Top languages used in repository card

`http://github-profile-summary-cards.vercel.app/api/cards/repos-per-language?username={username}&theme={theme_name}&exclude={exclude}`
- Accept url parameters
- theme
- Theme name
- username
- Username
- exclude:
- A comma separated list of languages to exclude, e.g., exclude=java,rust
- You can found the supported languages in [here](https://github.com/github/linguist/blob/master/lib/linguist/languages.yml)
### Top languages in commits card

`http://github-profile-summary-cards.vercel.app/api/cards/most-commit-language?username={username}&theme={theme_name}&exclude={exclude}`
- Accept url parameters
- theme
- Theme name
- username
- Username
- exclude:
- A comma separated list of languages to exclude, e.g., exclude=java,rust
- You can found the supported languages in [here](https://github.com/github/linguist/blob/master/lib/linguist/languages.yml)
### GitHub stats card

`http://github-profile-summary-cards.vercel.app/api/cards/stats?username={username}&theme={theme_name}`
- Accept url parameters
- theme
- Theme name
- username
- Username
### Productive time card

`http://github-profile-summary-cards.vercel.app/api/cards/productive-time?username={username}&theme={theme_name}&utcOffset={utcOffset}`
- accept url parameters
- theme
- username
- utcOffset
---
## How to use (GitHub Actions)
This action generate your github profile summary cards and make a commit to your repo.
You can also trigger action by yourself after add this action.
:star: [Follow tutorial](https://github.com/vn7n24fzkq/github-profile-summary-cards/wiki/Toturial) ( Recommendation ) :star:
#### First step
- You need create a [Personal access token](https://docs.github.com/en/github/authenticating-to-github/creating-a-personal-access-token) with correct permissions.
[Personal token permissions](https://github.com/vn7n24fzkq/github-profile-summary-cards/wiki/Personal-access-token-permissions)
- Add personal access token to repo secret.
#### Use template ( create a repository )
- [github-profile-summary-cards-example](https://github.com/vn7n24fzkq/github-profile-summary-cards-example)
- Action already setup in this template, you just need click `use this template button` to create your profile readme.
- After replace GITHUB_TOKEN with your repo secret and trigger action you can use everything in `profile-summary-card-output` folder.
#### Add to exist repository
- Add this action to repo and replace GITHUB_TOKEN in action yml file with your repo secret.
---
## GitHub Actions usage
After the action finished. You can see all of summary cards are in folder which named `profile-summary-card-output`.
`Note: Some summary cards might not be updated in time, because github raw file has cache time.`
```yml
name: GitHub-Profile-Summary-Cards
on:
schedule: # execute every 24 hours
- cron: "* */24 * * *"
workflow_dispatch:
jobs:
build:
runs-on: ubuntu-latest
name: generate-github-profile-summary-cards
permissions:
contents: write
steps:
- uses: actions/checkout@v3
- uses: vn7n24fzkq/github-profile-summary-cards@release
env: # default use ${{ secrets.SUMMARY_GITHUB_TOKEN }}, you should replace with your personal access token
GITHUB_TOKEN: ${{ secrets.SUMMARY_GITHUB_TOKEN }}
with:
USERNAME: ${{ github.repository_owner }}
# BRANCH_NAME is optional, default to main, branch name to push cards
BRANCH_NAME: "main"
# UTC_OFFSET is optional, default to zero
UTC_OFFSET: 8
# EXCLUDE is an optional comma seperated list of languages to exclude, defaults to ""
EXCLUDE: ""
# AUTO_PUSH is optional, a boolean variable default to true, whether automatically push generated files to desired branch
AUTO_PUSH: true
```
---
## Local Run
- Require `node 16`, lower versions should get some problems.
- Add personal access token to `.env` file. ex: `GITHUB_TOKEN=abcda69ddf66ae95538c5b1666591b59b4abc73a`
- Remember `npm run build` after modifying any code
```sh
npm run run [username] [UTC offset]
```
Example
```sh
npm run run vn7n24fzkq 8
```
- To locally run the API you can use the vercel dev package
```sh
vercel dev
```
## Deploy your own API on Vercel
Quickly deploy your own version!
[](https://vercel.com/new/clone?repository-url=https%3A%2F%2Fgithub.com%2Fvn7n24fzkq%2Fgithub-profile-summary-cards&env=GITHUB_TOKEN&envDescription=https%3A%2F%2Fgithub.com%2Fvn7n24fzkq%2Fgithub-profile-summary-cards%23first-step&project-name=my-github-profile-summary-cards)
|
2024-11-18T17:58:59.588733+00:00 | 2020-07-01T08:13:11 | ff344b6389d1057348fe0cecc2c4eb3f2b67412d | 2 | {
"blob_id": "ff344b6389d1057348fe0cecc2c4eb3f2b67412d",
"branch_name": "refs/heads/master",
"committer_date": "2020-07-01T08:13:11",
"content_id": "950feaf45ccd56300d1803379540f82c33b53fde",
"detected_licenses": [
"MIT"
],
"directory_id": "48981be0f9b3695234701773000306d73cf84424",
"extension": "md",
"filename": "classtime.md",
"fork_events_count": 2,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": 242881584,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 1506,
"license": "MIT",
"license_type": "permissive",
"path": "/Week1/classtime.md",
"provenance": "stack-edu-markdown-0000.json.gz:85",
"repo_name": "Corrina-Busby/WEB701_Journal",
"revision_date": "2020-07-01T08:13:11",
"revision_id": "46e2d48b6e0507e2eb47f5aa7a3be2d66a6aa61a",
"snapshot_id": "156ae4e71ecbf8fd3980afabe1b8ff948a7dd5c9",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/Corrina-Busby/WEB701_Journal/46e2d48b6e0507e2eb47f5aa7a3be2d66a6aa61a/Week1/classtime.md",
"visit_date": "2021-01-15T04:44:51.666321"
} | 2.078125 | stackv2 | # Course Overview
## Session 1
In class, Todd our level 7 WEB tutor had us choose our own group and put together a presentation centered on the requirements for all of our assignments for the semester.
For our presentation my main input covered assignment 2 Milestone Two and Assignment 3 Part 2. Performing this execise will hopefully help keep me focused on our tasks and the marking schedule and not become side tracked and overwhelmed with unnecessary information.
Our group "Team Middle Desk" have a shared presentation slide that I will need to refer to and read in my own time to understand what is required for the other assignments and milestones starting with assignment 2.
Assignment one is the mandatory Journal worth 10%, this should be some easy marks to attain
## Session 2
Information architecture
http://maya.com/blog/what-is-information-architecture
Two really good videos in understanding I.A
Information is about how we relay the message
Architecture is building from a template
Followed by walking through a student's brief from 2016 to give us an template of what is required in the first assignment
## Homework
No shortage of links to sieve about Information Architecture, a few Lynda tutorials, Express.js and node.js
Setup a git journal and send link to tutor
Fill out the brief template choosing a collective based around auctioning simillar to trade me and we have been specifically informed not to choose "Berries", therefore I am going with cows glorious evil cows
|
2024-11-18T18:04:28.955588+00:00 | 2020-09-06T20:39:05 | 66fd52ca33093b32f4a85d32551c23b2ce3a6c0b | 4 | {
"blob_id": "66fd52ca33093b32f4a85d32551c23b2ce3a6c0b",
"branch_name": "refs/heads/master",
"committer_date": "2020-09-06T20:39:05",
"content_id": "6e9617579f530f50120e6db265cb06d1ea827c5b",
"detected_licenses": [
"MIT"
],
"directory_id": "035cb8c4f7339b7e0492bd46b99c07ed177e8d67",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": 168756947,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 2564,
"license": "MIT",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:86",
"repo_name": "JoelLinn/haproxy-acme-validation-proxy-plugin",
"revision_date": "2020-09-06T20:39:05",
"revision_id": "713c71f70abb7535e57eb9eab8e9a0472c11be2f",
"snapshot_id": "394c92e7ffa59144d03c92f98da3fb92ed8410c5",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/JoelLinn/haproxy-acme-validation-proxy-plugin/713c71f70abb7535e57eb9eab8e9a0472c11be2f/README.md",
"visit_date": "2021-07-03T17:30:47.501883"
} | 3.671875 | stackv2 | # Deprecated
See [JoelLinn/acme-proxy](https://github.com/JoelLinn/acme-proxy) instead.
## `HAProxy` ACME domain validation plugin
[]()
[release]: https://github.com/janeczku/haproxy-acme-validation-plugin/releases
HAProxy plugin implementing [ACME http-01](https://github.com/ietf-wg-acme/acme/) validation for split dns domains isolated to the internet by HAProxy instances. The plugin leverages HAProxy's Lua API to allow HAProxy to forward validation challenges using token/key-auth provisioned by an ACME client on the intranet over http.
## Compatible ACME clients
The plugin is compatible with all ACME clients supporting http-01 challenges.
## Lightweight
Can run in an existing haproxy instance, only one additional lua file, lua-http and lua-sockets are needed.
## Installation instructions
### Prerequesites
You need to be rolling HAProxy version `1.6.0` or later with Lua support enabled.
To check if your HAProxy binary was compiled with Lua support run the following command:
haproxy -vv
If there is a line similar to this you are good to go:
Built with Lua version
If your binary doesn't come with Lua bindings, you can download Debian/Ubuntu packages of the latest v1.6 release from the [Debian HAProxy packaging team](http://haproxy.debian.net/).
### Installation
Copy `acme-http01-webroot.lua` to a location accessible by HAProxy.
To activate the plugin you just need to add **three lines** to your `haproxy.cfg`:
In the `global` section insert
lua-load /etc/haproxy/acme-http01-webroot.lua
to invoke the Lua plugin.
In the `frontend` section serving the domain(s) for which you want to create/renew certificates insert:
acl url_acme_http01 path_beg /.well-known/acme-challenge/
http-request use-service lua.acme-http01 if METH_GET url_acme_http01
This will pass ACME http-01 validation requests to the Lua plugin handler.
*Note:* ACME protocol stipulates validation on port 80. If your HTTP frontend listens on a non-standard port, make sure to add a port 80 bind directive.
Finally, soft-restart HAProxy (see below for instructions) to apply the updated configuration.
## Getting certificates
Make sure your desired domain globally (on the internet) resolves to haproxy or a firewall/gateway that forwards port 80 to your haproxy instance.
The system haproxy runs on needs to be configured to use the local dns so it forwards all challenge requests to machines on the intranet and not itself.
|
2024-11-18T18:04:29.006636+00:00 | 2020-01-07T10:50:50 | 5a3517cc2e51a7203dc6c85a66f266b66ba65a1a | 3 | {
"blob_id": "5a3517cc2e51a7203dc6c85a66f266b66ba65a1a",
"branch_name": "refs/heads/master",
"committer_date": "2020-01-07T10:50:50",
"content_id": "aa977f049e4ac9dc76b5879001f222b06f4f7853",
"detected_licenses": [
"MIT"
],
"directory_id": "18ca26ab3b2e54731fd3b535e93efbc241f33467",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": 232303857,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 590,
"license": "MIT",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:87",
"repo_name": "damir-majer/ABAPDevDirectory",
"revision_date": "2020-01-07T10:50:50",
"revision_id": "6879f3e364f9dcf36eafdc023b5a8bc589331c20",
"snapshot_id": "b57e12c783c65d0d6573c09733a3b7c212ea366d",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/damir-majer/ABAPDevDirectory/6879f3e364f9dcf36eafdc023b5a8bc589331c20/README.md",
"visit_date": "2020-12-06T01:36:15.544575"
} | 2.515625 | stackv2 | # ABAPDevDirectory
The ABAP Dev Directory
# iOS Dev Directory
This repository contains the Data for ABAP Dev Directory.
## Contributing your site
If your site covers ABAP development and you'd like to add it to the directory, or if you'd like to add someone else's relevant site, you're more than welcome to!
## What is the ABAP Dev Directory?
It's just a site that lists all of the blogs that cover the wonderful ABAP development community. It was built, and is maintained by [Damir Majer](https://twitter.com/majcon) who is also the author of [ABAP Weekly](https://abapweekly.com).
|
2024-11-18T18:04:29.462557+00:00 | 2022-10-18T13:08:32 | 1b48f5446777ef054b0cd997c84b20c4715e205d | 4 | {
"blob_id": "1b48f5446777ef054b0cd997c84b20c4715e205d",
"branch_name": "refs/heads/master",
"committer_date": "2022-10-18T13:08:32",
"content_id": "497e9d223753e2c2532dac68950fdef6497b92ad",
"detected_licenses": [
"MIT"
],
"directory_id": "f58b884ed4471549838347c649f69244020e3c9c",
"extension": "md",
"filename": "readme.md",
"fork_events_count": 1,
"gha_created_at": "2017-02-06T13:37:42",
"gha_event_created_at": "2022-10-18T13:08:33",
"gha_language": "PHP",
"gha_license_id": "MIT",
"github_id": 81090633,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 12760,
"license": "MIT",
"license_type": "permissive",
"path": "/readme.md",
"provenance": "stack-edu-markdown-0000.json.gz:88",
"repo_name": "konsulting/laravel-transformer",
"revision_date": "2022-10-18T13:08:32",
"revision_id": "3ebef4491b74c2a67059bc18318d18c2b5b4ac5f",
"snapshot_id": "c2b99c6a8a0203ae4bb9dc49e36b21b5bd1e99ef",
"src_encoding": "UTF-8",
"star_events_count": 12,
"url": "https://raw.githubusercontent.com/konsulting/laravel-transformer/3ebef4491b74c2a67059bc18318d18c2b5b4ac5f/readme.md",
"visit_date": "2023-01-28T18:02:38.896304"
} | 4.0625 | stackv2 | # Transformer
[](https://travis-ci.org/konsulting/laravel-transformer)
[](https://scrutinizer-ci.com/g/konsulting/laravel-transformer/?branch=master)
*A simple way to transform data, with helpers for transforming data in [Laravel](https://laravel.com) applications.*
## Installation
* Install Transformer using composer: `composer require konsulting/laravel-transformer`
### Transformer in a Laravel application
* Transformer uses `Illuminate\Support\Collection` and `Illuminate\Support\Arr`, and requires a couple of extensions to these.
The extensions are available in the `konsulting/laravel-extend-collections` package.
* You'll need both the `CollectionsServiceProvider` from that package, and `TransformerServiceProvider` in your `config/app.php`.
If you are using Laravel 5.5, they will auto-register. However, if you have chosen not to auto-register,
or are using an earlier version, add the following to `config/app.php`
```php
'providers' => [
// Other service providers...
Konsulting\Laravel\CollectionsServiceProvider::class,
Konsulting\Laravel\Transformer\TransformerServiceProvider::class,
],
```
* Optionally, add the Transformer Facade to `config/app.php`
```php
'aliases' => [
// Other aliases...
'Transformer' => Konsulting\Laravel\Transformer\TransformerFacade::class,
],
```
* Optionally publish the config file, and adjust the rule packs you want to use.
`php artisan vendor:publish --provider=Konsulting\\Laravel\\Transformer\\TransformerServiceProvider --tag=config`
### Transformer outside a Laravel application
* Transformer uses `Illuminate\Support\Collection` and `Illuminate\Support\Arr`. Outside a Laravel application,
it will use `tighten/collect` (an extraction of Collection && Arr from Laravel's Illuminate\Support) to get these dependencies.
* Transform also requires a couple of extensions to these. The extensions are available in the
`konsulting/laravel-extend-collections` package. You'll need to register the extensions manually.
* You will need to build up your Transformer manually for use in your application.
```php
// Basic example
use Konsulting\Laravel\Transformer\Transformer;
use Konsulting\Laravel\Transformer\RulePacks\CoreRulePack;
use Konsulting\Laravel\Transformer\RulePacks\CarbonRulePack;
require __DIR__ . '/../vendor/autoload.php';
// Extend Illuminate\Support\Arr and Illuminate\Support\Collection
\Konsulting\Laravel\load_collection_extensions();
// Build up Transformer
$transformer = new Transformer([CoreRulePack::class, CarbonRulePack::class]);
// Transformer now available to use, see Usage
```
## Usage
Transformer uses `RulePacks` to provide transformation functionality. RulePacks can be added to the Transformer during construction, or after with the `addRulePack` or `addRulePacks` methods.
A set of rules can be passed in during construction (useful when applying the same rules to different sets of data) or rules can be passed in at the point when performing transformation.
To transform data, the `transform` method is used. It accepts an array (or collection) of data to transform, and optionally rules to apply.
* Rules are presented in a similar manner to the [Laravel Validator](https://laravel.com/docs/5.4/validation). They provide functionality to handle nested data, and follow the same string format.
* Arrays of rules are indexed by a field expression and provide a `|` (pipe) delimited list of rules to apply.
* Rules may be provided a set of parameters in CSV format. Field expressions may use `*` as a wildcard to match elements at that depth and `**` as a special case to match everything.
* Rule sequences are built up in the order they are provided.
```php
// using the $transformer built up earlier
$rules = [
'*' => 'drop_if_empty',
'name' => 'trim',
'contacts.*.name' => 'trim|uppercase'
];
$data = [
0 => '',
'name' => ' Keoghan Litchfield ',
'contacts' => [
['name' => 'Robin'],
['name' => 'Roger'],
['name' => ''],
],
];
$result = $transformer->transform($data, $rules);
// Outputs [
// 'name' => 'Keoghan Litchfield',
// 'contacts' => [
// ['name' => 'ROBIN'],
// ['name' => 'ROGER'],
// ],
// ];
```
### Transform helper
There is also a helper class `Transform`, which facilitates the easy transformation of a single value by one or more rules. `Transform` receives an instance of `Transformer` via its constructor, which provides the transformation logic and determines which rules are available.
Using the instance of `Transformer` built up previously:
```php
use Konsulting\Laravel\Transformer\Transform;
$transform = new Transform($transformer);
```
Rules may be called as methods on the `Transform` object, with the value to be transformed passed in as the first argument and any rule parameters as subsequent arguments.
```php
$transform->trim(' Some string to be trimmed '); // Outputs 'Some string to be trimmed'
$transform->regexReplace('testing', 'e', 'oa'); // Outputs 'toasting'
```
Alternatively, rules may be passed via the `withRule()` and `withRules()` methods (for singular and multiple rules respectively).
Rule parameters are passed either as separate arguments, or as an array.
```php
// Single rule
$transform->withRule(' test ', 'trim'); // Outputs 'test'
// Single rule with parameters passed as separate arguments
$transform->withRule('test', 'regex_replace', 'e', 'oa'); // Outputs 'toast'
// Singe rule with parameters passed as an array
$transform->withRule('test', 'regex_replace', ['e', 'oa']); // Outputs 'toast' as well
// Multiple rules passed as a sequential array
$transform->withRules(' test ', ['trim', 'uppercase']); // Outputs 'TEST'
// Multiple rules and parameters passed as an assocative array: [$rule => [$param1, $param2], $rule2 => []...]
$transform->withRules('--test--', [ // Outputs 'TOAST'
'trim' => ['-'],
'regex_replace' => ['e', 'oa'],
'uppercase' => [],
]);
```
#### Fluent API
Rules may also be called fluently: the input value is set with the `input()` method, and the result is obtained with `get()`.
Any number of rule methods may be chained between these.
```php
$transform->input(' hello ')
->trim()
->regexReplace('hello', 'world')
->uppercase()
->get();
// Outputs 'WORLD'
```
When the fluent API is used, the value is not passed as an argument to the rule methods (as it has already been set via `input()`).
As such, all arguments passed to rule methods are treated as rule parameters.
`withRule()` and `withRules()` may be used to fluently declare rules with or without parameters:
```php
$transform->input($input)
->withRule('trim')
->uppercase()
->get();
$transform->input($input)
->lowercase()
->withRules(['trim', 'uppercase'])
->get();
```
### Available Rules
We provide a couple of rule packs for use, it is easy to extend the rules available by creating your own Rule Pack. Rule Packs are loaded in the declared order, methods in later packs will override packs loaded earlier.
Parameter names are denoted by `<param>` and optional parameters by `[<param>]`.
#### Core Rule Pack
##### Cleaning up
* `null_if_empty`
* `null_if_empty_string`
* `return_null_if_empty` – equivalent to `null_if_empty|bail_if_null`.
* `return_null_if_empty_string` – equivalent to `null_if_empty_string|bail_if_null`.
* `bail_if_null`
* `drop_if_null`
* `drop_if_empty`
* `drop_if_empty_string`
* `trim:[<character(s)_to_trim>]` – performs default PHP trim() if no characters supplied.
##### Casting
* `string` - convert to string, an array is transformed to a CSV or returns ‘’ for items that cannot be represented as a string.
* `boolean`
* `array`
* `collection` - convert to `Illuminate\Support\Collection`
* `json`
* `float`
* `integer`
* `date_time:[timezone]`
* `date_time_immutable:[timezone]`
##### String manipulation
* `uppercase`
* `lowercase`
##### Regex and string replace
* `replace:<search_string>,<replace_string>`
* `regex_replace:<search_regex>,<replace_string>`
* `numeric`
* `alpha`
* `alpha_dash`
* `alpha_num`
* `alpha_num_dash`
#### Carbon Rule Pack
* `carbon`
* `date_format` - parameter for the format required.
#### Related Fields Rule Pack
* `null_with:<key>`
* `drop_with:<key>`
* `bail_with:<key>`
* `null_without:<key>`
* `drop without:<key>`
* `bail_without:<key>`
#### Number Rule Pack
* `clamp:<min>,<max>` - constrain a number between two values
_**Note:** key is a dot notation key for another field in the dataset_
### Laravel Helpers
We use Laravel frequently, so have a couple of extras added here.
#### Facade
Use the facade to gain easy access to the Transformer wherever you are.
```php
\Transformer::transform($data, $rules);
```
#### Request Macro `transform`
The Service Provider adds the transform Macro to the `Illuminate\Http\Request` class. This makes it simple to invoke the transformation on a request at any point. The method passes the request object back to allow chaining.
```php
// Example Controller
namespace App\Http\Controllers;
use App\ContactRequest;
use Illuminate\Http\Request;
class ContactRequestsController
{
// ...
public function store(Request $request)
{
$request->transform([
'name' => 'trim|uppercase',
'message' => 'trim',
]);
$this->validate($request, [
'name' => 'required',
]);
return ContactRequest::create(
$request->only('name', 'message')
);
}
}
```
#### TransformingRequest Trait
This trait makes the form request transform the data before validation occurs (which is upon it being resolved by the container).
Rules for transformation are provided in the `transformRules` method.
Using the TransformingRequest Trait
```php
// Form Request
namespace App\Http\Requests;
use Illuminate\Foundation\Http\FormRequest;
use Konsulting\Laravel\Transformer\TransformingRequest;
class ContactFormRequest extends FormRequest
{
use TransformingRequest;
// ... other methods including rules()
public function transformRules()
{
return [
'name' => 'trim|uppercase',
'message' => 'trim',
];
}
```
```php
// Controller
namespace App\Http\Controllers;
use App\ContactRequest;
use App\Http\Requests\ContactFormRequests;
class ContactRequestsController
{
// ...
public function store(ContactFormRequest $request)
{
return ContactRequest::create(
$request->only('name', 'message')
);
}
}
```
### Middleware
The `TransformRequest` middleware applies transformations to requests according to configured rules. These rules are specified in the `middleware_rules` key of the config file as detailed in [Usage](#usage).
To register the middleware for use in your project, add the following line to your project's `App/Http/Kernel.php`:
```php
'transform_data' => \Konsulting\Laravel\Transformer\Middleware\TransformRequest::class
```
The default middleware rules state that every field should be trimmed of whitespace and nulled if empty:
```php
'middleware_rules' => [
'**' => 'trim|return_null_if_empty',
]
```
Rules need not be applied to all fields; specific fields may be targeted within the middleware if required:
```php
'middleware_rules' => [
'postcode' => 'uppercase',
'email' => 'lowercase',
]
```
With the above configuration, the postcode and email fields of every request sent through the middleware will be affected, but all other fields will be left unchanged.
Multiple transformer middlewares may be useful in a project: to achieve this, copy `laravel-transformer/src/Middleware/TransformRequest.php` to your project's `App/Http/Middleware` directory, and rename/edit as necessary. Each new middleware will have to be registered in the kernel.
## Contributing
Contributions are welcome and will be fully credited. We will accept contributions by Pull Request.
Please:
* Use the PSR-2 Coding Standard.
* Add tests, if you’re not sure how, please ask.
* Document changes in behaviour, including readme.md.
## Testing
We use [PHPUnit](https://phpunit.de) and the excellent [orchestral/testbench](https://github.com/orchestral/testbench).
Run tests using PHPUnit: `vendor/bin/phpunit`
|
2024-11-18T18:04:29.508786+00:00 | 2020-02-19T03:28:12 | 1ff7f6c7cfb083578bb15c7c5ca1e53fb70634a4 | 3 | {
"blob_id": "1ff7f6c7cfb083578bb15c7c5ca1e53fb70634a4",
"branch_name": "refs/heads/master",
"committer_date": "2020-02-19T03:28:12",
"content_id": "01543fc17b2f41d9ca7bd89e9f919c8466cfcfc4",
"detected_licenses": [
"MIT"
],
"directory_id": "9d9adf609d7f0b8aed346c88a5df729cd372fbb5",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": 229834089,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 1325,
"license": "MIT",
"license_type": "permissive",
"path": "/ZX81QuickLook/ZX81 Functions/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:89",
"repo_name": "sebastienboisvert/ZX81QuickLook",
"revision_date": "2020-02-19T03:28:12",
"revision_id": "ffb278e9d37f81ffda57dd8f17daea9b74e61eb0",
"snapshot_id": "db7e0be91abe6e39b7c6f0c5eb407c87d1fba503",
"src_encoding": "UTF-8",
"star_events_count": 4,
"url": "https://raw.githubusercontent.com/sebastienboisvert/ZX81QuickLook/ffb278e9d37f81ffda57dd8f17daea9b74e61eb0/ZX81QuickLook/ZX81 Functions/README.md",
"visit_date": "2022-04-02T20:28:01.957932"
} | 3.015625 | stackv2 | # ZX81QuickLook Functions
The main processing functions are found here. Most of the functions take a `FILE` pointer of the .P file being processed, and a `string_buffer`
struct with a character buffer that will contain the processed data. In-memory buffers are used to avoid creating temporary files for the data, since Quicklook imposes some limits on access to external resources for HTML display.
The main functions are:
- `checkFile()`: this does some sanity checks on the file, and returns a `ZX81_Program` struct that contains various pieces of info used by other
functions
- `get_basic_listing()`: this produces the basic listing
- `get_screen()`: generates the screen output, then checks if there are actual printable characters to determine if the interface should show that tab
- `get_save_sound()`: generates WAV data for the save audio for playback. This uses the `p2raw.c` functions found in `ThirdParty/` for generating
the raw audio. For this function, in-memory `FILE` objects are used since the original code operated on files directly.
These functions are self-contained and should be easy to use independently (no dependencies to other project code - both the CLI `ZX81QL` scheme and the main Quicklook plugin schemes share these functions). All the code is commented to (hopefully) be easy to follow.
|
2024-11-18T18:04:29.822701+00:00 | 2023-08-04T13:13:10 | 9b2a8bc2bf19bcf49769d7918ff1d1b87541ed68 | 3 | {
"blob_id": "9b2a8bc2bf19bcf49769d7918ff1d1b87541ed68",
"branch_name": "refs/heads/master",
"committer_date": "2023-08-04T13:13:10",
"content_id": "2c8b9a1608d76f06f52dcdad490448477313ba74",
"detected_licenses": [
"Apache-2.0"
],
"directory_id": "4491549f0b1bbf5397ae0b56192605a7abcb61b0",
"extension": "md",
"filename": "SECURITY.md",
"fork_events_count": 0,
"gha_created_at": "2020-01-21T15:33:01",
"gha_event_created_at": "2023-08-05T00:31:43",
"gha_language": "Python",
"gha_license_id": "Apache-2.0",
"github_id": 235376119,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 984,
"license": "Apache-2.0",
"license_type": "permissive",
"path": "/SECURITY.md",
"provenance": "stack-edu-markdown-0000.json.gz:90",
"repo_name": "iNoSec2/skf-labs",
"revision_date": "2023-08-04T13:13:10",
"revision_id": "8af9edc83e313be1578c5dee0fd4ecdf7ac18a32",
"snapshot_id": "81e9d400ccac1007632add23bd50a094de1f50d5",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/iNoSec2/skf-labs/8af9edc83e313be1578c5dee0fd4ecdf7ac18a32/SECURITY.md",
"visit_date": "2023-08-17T00:20:12.274684"
} | 2.734375 | stackv2 | # Security Policy
## Reporting a Vulnerability
The OWASP SKF Labs are intentionally vulnerable web applications, provided for educational purporses.
Containerized applications that include security issues are most likely built that way on purpose, for students to attack and exploit.
However, it is never our intention to cause potential harm to the students' computer systems nor to other resources on the network! If you feel that a container, or other part of our project, can cause potential harm to your system(s), please [raise an issue through our Github project page](https://github.com/blabla1337/skf-labs/issues).
While we are not a highly active group, we certainly strive to deliver a safe and educational experience. Any security issues in our products will be addressed as soon as possible. You should get a respons to a reported security issue within two weeks (ten working days).
You are of course very welcome to offer suggested fixes through a pull request!
|
2024-11-18T18:04:29.863489+00:00 | 2020-10-31T01:36:31 | c4957dcc66306ee4cf912d5f48a3ad438822fca9 | 3 | {
"blob_id": "c4957dcc66306ee4cf912d5f48a3ad438822fca9",
"branch_name": "refs/heads/main",
"committer_date": "2020-10-31T01:36:31",
"content_id": "86326a3a7e432dd5232540170ac3864cbd85e8a8",
"detected_licenses": [
"Apache-2.0"
],
"directory_id": "c1ae31bbe0531c3ece8a3f69e0dacc51b87b0f95",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": 307852911,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 920,
"license": "Apache-2.0",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:91",
"repo_name": "drinkcoffee/abi2",
"revision_date": "2020-10-31T01:36:31",
"revision_id": "396196ca24152324d51fd008989ff6707f0c15ba",
"snapshot_id": "ed1c6a805e668542197fb1a0a8eccaf066c7082d",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/drinkcoffee/abi2/396196ca24152324d51fd008989ff6707f0c15ba/README.md",
"visit_date": "2023-01-03T07:31:26.682526"
} | 2.8125 | stackv2 | # ABI2 Encoding Check
This repo contains code to check the support for ABI2 by Web3J and Solidity.
The directories:
* check: Solidity and Java code.
* gradle: build system related files.
* scripts: Scripts to set-up and run Hyperledger Besu.
## Building
This repo relies on a special version of Web3J. To create this:
* Clone the repo to the directory ./LTACFC/.. : git clone https://github.com/drinkcoffee/web3j-rlp
* cd ../web3j-rlp
* ./gradlew build
* cd codegen/build/distributions
* tar xvf codegen-4.7.0-SNAPSHOT.tar
The runner in check (abi2/check/src/main/java/check/Main.java) relies on a properties file of the form:
PRIVATE_KEY=40000000000ad044fb60865556ead3961456349a242b3a4b8a00000000000001
//ROOT_BC_ID=4
//ROOT_URI=https://rinkeby.infura.io/v3/4000000000000000000000124e8fae13
//ROOT_GAS=HIGHEST
//ROOT_PERIOD=15000
ROOT_BC_ID=1F
ROOT_URI=http://<IP_ADDRESS>:8310/
ROOT_GAS=FREE
ROOT_PERIOD=2000
|
2024-11-18T18:04:29.907619+00:00 | 2021-02-25T18:19:33 | 28cb0a74be1bb016ddb03953dc3f587b782f1bb8 | 2 | {
"blob_id": "28cb0a74be1bb016ddb03953dc3f587b782f1bb8",
"branch_name": "refs/heads/main",
"committer_date": "2021-02-25T18:19:33",
"content_id": "7787cff101fda2a6c00b1ab9b2caef44ff3d5745",
"detected_licenses": [
"MIT"
],
"directory_id": "0b5929af7c035ebe70d7f9f549a4bd0f8e90ddd3",
"extension": "md",
"filename": "CHANGELOG.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": null,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 2442,
"license": "MIT",
"license_type": "permissive",
"path": "/CHANGELOG.md",
"provenance": "stack-edu-markdown-0000.json.gz:92",
"repo_name": "isabella232/semantic-release-config-logdna",
"revision_date": "2021-02-25T18:19:33",
"revision_id": "81827945b61df62a43d0eb6194b6aa2101d6343d",
"snapshot_id": "c35840bce83c20fc0c3d3e920625757dc977bed4",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/isabella232/semantic-release-config-logdna/81827945b61df62a43d0eb6194b6aa2101d6343d/CHANGELOG.md",
"visit_date": "2023-03-07T12:10:36.721870"
} | 2.421875 | stackv2 | ## Changelog
## [1.1.1](https://github.com/logdna/semantic-release-config-logdna/compare/v1.1.0...v1.1.1) (2021-02-25)
### Bug Fixes
* **rules**: map any breaking changes to a major version [bd2b31a](https://github.com/logdna/semantic-release-config-logdna/commit/bd2b31a687a8db4f40ef4711b969918bdad31954) - Eric Satterwhite
# [1.1.0](https://github.com/logdna/semantic-release-config-logdna/compare/v1.0.2...v1.1.0) (2021-02-11)
### Features
* **package**: add mike del tito as a contributor [836f4f8](https://github.com/logdna/semantic-release-config-logdna/commit/836f4f8ca9da20eed9bbf1b15454b9628a4fca20) - Eric Satterwhite
### Miscellaneous
* add @esatterwhite as a contributor [d3268b4](https://github.com/logdna/semantic-release-config-logdna/commit/d3268b4a67f77d415d814a46e3656b4bac9f3882) - Eric Satterwhite
* add @mdeltito as a contributor [4a8f6b9](https://github.com/logdna/semantic-release-config-logdna/commit/4a8f6b9f4c8a94dacd7a554e42cdbc7270665eea) - Eric Satterwhite
## [1.0.2](https://github.com/logdna/semantic-release-config-logdna/compare/v1.0.1...v1.0.2) (2021-02-11)
### Bug Fixes
* **commit**: the bang (!) indicates a breaking change [034b30d](https://github.com/logdna/semantic-release-config-logdna/commit/034b30d00ee5a00d246a005485af4184031112a7) - Eric Satterwhite, closes: [#5](https://github.com/logdna/semantic-release-config-logdna/issues/5)
## [1.0.1](https://github.com/logdna/semantic-release-config-logdna/compare/v1.0.0...v1.0.1) (2021-02-10)
### Bug Fixes
* update package name globally [a9598fa](https://github.com/logdna/semantic-release-config-logdna/commit/a9598fa31b1752d35721472b1909ac0762052c17) - Mike Del Tito
### Documentation
* update installation instructions [0aba75a](https://github.com/logdna/semantic-release-config-logdna/commit/0aba75a017da013e2f8a0cec2c0c92065add3a1a) - Mike Del Tito
# 1.0.0 (2021-02-09)
### Bug Fixes
* **ci**: fix incorrect reference to npm credentials [7ab073c](https://github.com/logdna/semantic-release-config-logdna/commit/7ab073c193598ca0804a4ba2164bba20fc827811) - Eric Satterwhite
### Features
* **pkg**: initial release config [d07807b](https://github.com/logdna/semantic-release-config-logdna/commit/d07807b8bf02a5f40ebc6096190594eb8ea3b85b) - Eric Satterwhite
### Miscellaneous
* Initial commit [e435789](https://github.com/logdna/semantic-release-config-logdna/commit/e4357892c05f790968f1d63a3319173d02ca62c9) - GitHub
|
2024-11-18T18:04:29.969820+00:00 | 2016-12-06T15:34:55 | 7afb0b4dcf8d08a43b69ad30339a848f6b894277 | 2 | {
"blob_id": "7afb0b4dcf8d08a43b69ad30339a848f6b894277",
"branch_name": "refs/heads/master",
"committer_date": "2016-12-06T15:34:55",
"content_id": "813d603c8918411a69556d9120ad66e0d5d166c0",
"detected_licenses": [
"MIT"
],
"directory_id": "f36331166b9676e4242008524ea32f4033875232",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": 75743406,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 417,
"license": "MIT",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:93",
"repo_name": "chriskottom/rails5-api-blog-demo",
"revision_date": "2016-12-06T15:34:55",
"revision_id": "64d0206a4afb51a597f0f545930c4ec67ff8e401",
"snapshot_id": "b02d0eff5fce0f053e18de65f0526e0961b8aeaf",
"src_encoding": "UTF-8",
"star_events_count": 1,
"url": "https://raw.githubusercontent.com/chriskottom/rails5-api-blog-demo/64d0206a4afb51a597f0f545930c4ec67ff8e401/README.md",
"visit_date": "2021-06-07T18:05:28.314191"
} | 2.484375 | stackv2 | # Rails API Blog Demo
This is a demo application that uses a Rails 5 API application with
ActiveModel::Serializers to build a JSON:API-compatible interface.
## References
* Tutorial - https://www.simplify.ba/articles/2016/06/18/creating-rails5-api-only-application-following-jsonapi-specification/
* JSON:API - http://jsonapi.org/
* ActiveModel::Serializers - https://github.com/rails-api/active_model_serializers
|
2024-11-18T18:04:30.001171+00:00 | 2021-09-08T06:55:56 | f0383ed896e501e8f7b0863f5e0c096bee905d70 | 4 | {
"blob_id": "f0383ed896e501e8f7b0863f5e0c096bee905d70",
"branch_name": "refs/heads/master",
"committer_date": "2021-09-08T06:55:56",
"content_id": "3c3d2788baa5f958711e17fe176b1349ee46fc14",
"detected_licenses": [
"Apache-2.0"
],
"directory_id": "d3ae3bbd156039389cb31f1619d5fd4d234ee647",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": 404232278,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 2666,
"license": "Apache-2.0",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:94",
"repo_name": "umersheikh846/WQMIX",
"revision_date": "2021-09-08T06:55:56",
"revision_id": "3fbc088273c4ddb6876a352d18f057c50703546c",
"snapshot_id": "520e3b8b982f0816c7078990196e447bb83f6caa",
"src_encoding": "UTF-8",
"star_events_count": 5,
"url": "https://raw.githubusercontent.com/umersheikh846/WQMIX/3fbc088273c4ddb6876a352d18f057c50703546c/README.md",
"visit_date": "2023-07-17T21:39:46.221188"
} | 3.609375 | stackv2 | # Code of WQMIX used in the paper: Learning Fair Policies in Decentralized Cooperative Multi-Agent Reinforcement Learning (https://arxiv.org/abs/2012.09421)
Please use the pymarl\_wrapper for the environments used in our paper. Please see the instructions below for SMAC installation and WQMIX setup.
## Weighted QMIX: Expanding Monotonic Value Function Factorisation (NeurIPS 2020)
Based on PyMARL (https://github.com/oxwhirl/pymarl/). Please refer to that repo for more documentation.
This repo contains the cleaned-up code that was used in "Weighted QMIX: Expanding Monotonic Value Function Factorisation" (https://arxiv.org/abs/2006.10800).
## Included in this repo
In particular implementations for:
- OW-QMIX
- CW-QMIX
- Versions of DDPG & SAC used in the paper
We thank the authors of "QPLEX: Duplex Dueling Multi-Agent Q-Learning" (https://arxiv.org/abs/2008.01062) for their implementation of QPLEX (https://github.com/wjh720/QPLEX/), whose implementation we used. The exact implementation we used is included in this repo.
Note that in the repository the naming of certain hyper-parameters and concepts is a little different to the paper:
- α in the paper is `w` in the code
- Optimistic Weighting (OW) is referred to as `hysteretic_qmix`
## For all SMAC experiments we used SC<IP_ADDRESS>.69232 (not SC2.4.10). The underlying dynamics are sufficiently different that you **cannot** compare runs across the 2 versions!
The `install_sc2.sh` script will install SC<IP_ADDRESS>.69232.
## Running experiments
The config files (`src/config/algs/*.yaml`) contain default hyper-parameters for the respective algorithms.
These were changed when running the experiments for the paper (`epsilon_anneal_time = 1000000` for the robustness to exploration experiments, and `w=0.1` for the predator prey punishment experiments for instance).
Please see the Appendix of the paper for the exact hyper-parameters used.
Set `central_mixer=atten` to get the modified mixing network architecture that was used for the final experiment on `corridor` in the paper.
As an example, to run the OW-QMIX on 3s5z with epsilon annealed over 1mil timesteps using docker:
```shell
bash run.sh $GPU python3 src/main.py --config=ow_qmix --env-config=sc2 with env_args.map_name=3s5z w=0.5 epsilon_anneal_time=1000000
```
## Citing
Bibtex:
```
@inproceedings{rashid2020weighted,
title={Weighted QMIX: Expanding Monotonic Value Function Factorisation},
author={Rashid, Tabish and Farquhar, Gregory and Peng, Bei and Whiteson, Shimon},
booktitle={Advances in Neural Information Processing Systems},
year={2020}
}
```
## License
Code licensed under the Apache License v2.0
|
2024-11-18T18:04:30.095808+00:00 | 2022-02-20T23:03:41 | 057221694126e9258cfa78a65df2c3cc1242a2d6 | 3 | {
"blob_id": "057221694126e9258cfa78a65df2c3cc1242a2d6",
"branch_name": "refs/heads/master",
"committer_date": "2022-02-20T23:03:41",
"content_id": "eeb4ecd01f082f250acfae2b1d7532f8641debe0",
"detected_licenses": [
"MIT"
],
"directory_id": "4da3eca681d465f4994e84f43472cf66664e9b02",
"extension": "md",
"filename": "e6b38c6c75fdd1e5656e802f50e5ada24e664a2544793eadecae5a0c073e6185.md",
"fork_events_count": 2,
"gha_created_at": "2018-06-20T20:02:33",
"gha_event_created_at": "2023-04-12T05:47:06",
"gha_language": null,
"gha_license_id": "MIT",
"github_id": 138080116,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 2838,
"license": "MIT",
"license_type": "permissive",
"path": "/_emails/1996/10/e6b38c6c75fdd1e5656e802f50e5ada24e664a2544793eadecae5a0c073e6185.md",
"provenance": "stack-edu-markdown-0000.json.gz:95",
"repo_name": "cryptoanarchywiki/mailing-list-archive-generator",
"revision_date": "2022-02-20T23:03:41",
"revision_id": "5ee11c76b130aadf0ed74877107df5053ab0361b",
"snapshot_id": "7547cba9dd353f2f6dcc64eab77473ed5967c734",
"src_encoding": "UTF-8",
"star_events_count": 5,
"url": "https://raw.githubusercontent.com/cryptoanarchywiki/mailing-list-archive-generator/5ee11c76b130aadf0ed74877107df5053ab0361b/_emails/1996/10/e6b38c6c75fdd1e5656e802f50e5ada24e664a2544793eadecae5a0c073e6185.md",
"visit_date": "2023-04-16T03:14:42.652593"
} | 2.640625 | stackv2 | ---
layout: default
---
# 1996-10-14 - FLY_not
## Header Data
From: John Young<EMAIL_ADDRESS>To<EMAIL_ADDRESS>Message Hash: e6b38c6c75fdd1e5656e802f50e5ada24e664a2544793eadecae5a0c073e6185<br>
Message ID<EMAIL_ADDRESS>Reply To: _N/A_<br>
UTC Datetime: 1996-10-14 13:32:21 UTC<br>
Raw Date: Mon, 14 Oct 1996 06:32:21 -0700 (PDT)<br>
## Raw message
```
{% raw %}From: John Young<EMAIL_ADDRESS>Date: Mon, 14 Oct 1996 06:32:21 -0700 (PDT)
To<EMAIL_ADDRESS>Subject: FLY_not
Message-ID<EMAIL_ADDRESS>MIME-Version: 1.0
Content-Type: text/plain
10-13-96. CoWo:
"Encryption confusion"
Reminded of the secret Clipper algorithm, I sought
details from IBM. I asked a spokeswoman why the company
hadn't just put out a nice, snappy white paper
explaining its new approach to key recovery. "We spent
three months trying to do that, quite literally," the
spokeswoman said. "It's pretty confusing stuff, and
whenever we get it on paper, we aren't happy with it."
10-12-96. InWe:
"IBM Coalition May Not Fly" [Thanks, WR]
"The relaxation on encryption is in many ways an
extortion of industry," said Bob Gargus, president of
the Atalla. "If this starts to happen for export, how
many companies will be able to support two standards? I
think every American should be worried." But Greg Simon,
domestic policy adviser to Vice President Al Gore, said,
"We're not looking for an answer that's universally
popular. We're looking for a solution that's balanced
and fair."
10-11-96. RDSN:
"Bellcore Theory Affects Security"
"Future applications, some theorize, that work on smart
cards might work on your Pentium PC," says William Barr.
Most smart cards use private key encryption, such as DES
and RSA, making it all the more unlikely that Bellcore's
theory would work on smart cards. However, the threat
still should be investigated because the intention
eventually is to use public key encryption for wireless
communications and identification validation.
"ABA: Banks Should Issue Cards"
Only deposit institutions should issue smart cards, says
the American Bankers Association.
-----
http://jya.com/flynot.txt
ftp://jya.com/pub/incoming/flynot.txt
FLY_not
{% endraw %}
```
## Thread
+ Return to [October 1996](/archive/1996/10)
+ Return to "[John Young <jya<span>@</span>pipeline.com>](/authors/john_young_jya_at_pipeline_com_)"
+ 1996-10-14 (Mon, 14 Oct 1996 06:32:21 -0700 (PDT)) - FLY_not - _John Young<EMAIL_ADDRESS>
|
2024-11-18T18:04:30.154961+00:00 | 2021-10-28T08:06:42 | bf9e9540baf79a2606f23d75afd8e2088187ee04 | 4 | {
"blob_id": "bf9e9540baf79a2606f23d75afd8e2088187ee04",
"branch_name": "refs/heads/main",
"committer_date": "2021-10-28T08:06:42",
"content_id": "74d2132915b96ce024ac074151195a9547dcbfe8",
"detected_licenses": [
"Apache-2.0"
],
"directory_id": "d7397ff95fbec0d8f461ef56e04a653d7c222155",
"extension": "md",
"filename": "application_gitsync.md",
"fork_events_count": 0,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": null,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 1913,
"license": "Apache-2.0",
"license_type": "permissive",
"path": "/docs/resources/application_gitsync.md",
"provenance": "stack-edu-markdown-0000.json.gz:96",
"repo_name": "okamosy/terraform-provider-harness",
"revision_date": "2021-10-28T08:06:42",
"revision_id": "8edb2f51c479a7ebbc65e8586dbeee2bb5f134d8",
"snapshot_id": "d5c674d97f39668d2d62e1333aad6de4588c173b",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/okamosy/terraform-provider-harness/8edb2f51c479a7ebbc65e8586dbeee2bb5f134d8/docs/resources/application_gitsync.md",
"visit_date": "2023-08-23T05:43:08.758886"
} | 3.984375 | stackv2 | ---
# generated by https://github.com/hashicorp/terraform-plugin-docs
page_title: "harness_application_gitsync Resource - terraform-provider-harness"
subcategory: ""
description: |-
Resource for configuring application git sync.
---
# harness_application_gitsync (Resource)
Resource for configuring application git sync.
## Example Usage
```terraform
data "harness_secret_manager" "default" {
default = true
}
resource "harness_encrypted_text" "github_token" {
name = "github_token"
value = "<TOKEN>"
secret_manager_id = data.harness_secret_manager.default.id
}
resource "harness_git_connector" "myrepo" {
name = "myrepo"
url = "https://github.com/someorg/myrepo"
branch = "main"
generate_webhook_url = true
username = "someuser"
password_secret_id = harness_encrypted_text.github_token.id
url_type = "REPO"
}
resource "harness_application" "example" {
name = "example-app"
}
resource "harness_application_gitsync" "example" {
app_id = harness_application.example.id
connector_id = harness_git_connector.myrepo.id
branch = "main"
enabled = false
}
```
<!-- schema generated by tfplugindocs -->
## Schema
### Required
- **app_id** (String) The id of the application.
- **branch** (String) The branch of the git repository to sync to.
- **connector_id** (String) The id of the git connector to use.
### Optional
- **enabled** (Boolean) Whether or not to enable git sync.
- **id** (String) The ID of this resource.
- **repository_name** (String) The name of the git repository to sync to. This is only used if the git connector is for an account and not an individual repository.
## Import
Import is supported using the following syntax:
```shell
# Import using the Harness application id
terraform import harness_application_gitsync.myapp Xyz123
```
|
2024-11-18T18:04:30.303393+00:00 | 2017-05-05T16:47:17 | e11156a965fd879f697d8600e6e071fee72f6444 | 3 | {
"blob_id": "e11156a965fd879f697d8600e6e071fee72f6444",
"branch_name": "refs/heads/master",
"committer_date": "2017-05-05T16:47:17",
"content_id": "85a1aeeea4b296b95d38e0ff0e08bd55f42eb460",
"detected_licenses": [
"Apache-2.0"
],
"directory_id": "036bb8ff426882560e83114ff33c4970a0b0a0c6",
"extension": "md",
"filename": "README.md",
"fork_events_count": 0,
"gha_created_at": "2017-05-07T06:53:32",
"gha_event_created_at": "2017-05-07T06:53:32",
"gha_language": null,
"gha_license_id": null,
"github_id": 90514450,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 3429,
"license": "Apache-2.0",
"license_type": "permissive",
"path": "/README.md",
"provenance": "stack-edu-markdown-0000.json.gz:97",
"repo_name": "gitter-badger/silo",
"revision_date": "2017-05-05T16:47:17",
"revision_id": "6dfceb3071d1e7d94c194713e49f9fa776a9eee1",
"snapshot_id": "5669f7f4b4d53ebaf59fd6229c6408bf28b7d051",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/gitter-badger/silo/6dfceb3071d1e7d94c194713e49f9fa776a9eee1/README.md",
"visit_date": "2021-01-20T13:41:55.572088"
} | 3.203125 | stackv2 | # Summary
Silo represents a different take on container orchestration. At least, that's the intention, it's early days.
Below sets out the intended vision and the current plans to date. Silo and any associated repositories are open source, community driven development is our goal.
# What problem are we trying to solve?
Container's are totally awesome for many reasons, one of the keys to their success, is their simplicity and ease-of-use.
When building and developing on containers, at some stage you encounter container orchestration platforms such as Kubernetes and Rancher.
These systems are really, really powerful and are very appropriate to a lot of deployments. They represent a giant leap in the quality and feature sets over what came before.
Sometimes though, you just want a host to come online and run a pre-prescribed set of containers, perhaps as part of an auto-scaling group or similar.
When using services available in the cloud such as load balancers, this simple way of scaling compute in a predictable fashion, makes a lot of sense. You might be looking to keep things simple or maybe your looking to ensure consistent performance for peak periods.
Container orchestrators cater for this type of deployment, primarily though, such tools are built around the concept of handling the placement of containers themselve, provision of value-add components that wrap around your services.
In reality you can find yourself in such situations using a hugely powerful orchestration and scheduling tool to execute what should be a simple task. The ability for these platforms to handle encrypted networks, provide auto-discovery is excellent, but at times it can feel awkward when going against the grain of their typical use-cases.
*Surely there is a simpler way?*
You might then think that configuration management tools, Chef or Ansible might be a good option.
Whilst very capable of handling such types of deployment, they don't always fit the jigsaw quite so well with other container tools, nor do they give you good visibility of your services.
# Silo
So what's Silo and how does it fit in? Whilst not by any means a new concept, providing a simple and easy to use way to upon the boot, obtain a container run-list and start the containers. Instead of 'orchestrating', we focus on the most simple scheduling we can. This will be done by simply running a silo-agent container, servers can obtain their run-list and start the services on creation or boot.
That 'run-list' would be something remotely available to the server (S3, git, http, volume mount). At it's simplest form, this agent/run-list represent the 'runtime' part of the stack.
This gives people an easy way to stand-up containers on hosts. Sounds great... surely bash scripts or docker-compose could be used in such a fashion to achieve the same goals? This is where the server part of the equation brings benefit. The intention is to have a central server, fully independent from the runtime.
This optional server component would then be able to handle:
* Agent/Container State monitoring
* Metrics
* Ability to mark servers as down/degraded
* Integration with release/upgrade tooling
The features will likely be subject to a high rate of change in the initial stages. Fundamentally though, we want to create something that's small and simple. With the aim of reducing the overhead and complexity of running services this way.
|
2024-11-18T18:04:30.348605+00:00 | 2017-09-18T05:53:22 | 57e75f775990a280bbeb2e03ba708ec305a92af3 | 3 | {
"blob_id": "57e75f775990a280bbeb2e03ba708ec305a92af3",
"branch_name": "refs/heads/master",
"committer_date": "2017-09-18T05:53:22",
"content_id": "35c04b488f23a305c8d70e3714aec9db90092239",
"detected_licenses": [
"MIT"
],
"directory_id": "f61c38eafa0627ff34d51381df43e1fec7a10224",
"extension": "md",
"filename": "_index.md",
"fork_events_count": 0,
"gha_created_at": "2017-09-15T00:11:35",
"gha_event_created_at": "2017-09-15T00:11:35",
"gha_language": null,
"gha_license_id": null,
"github_id": 103594169,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 5592,
"license": "MIT",
"license_type": "permissive",
"path": "/web/content/title-8/chapter-92F/_index.md",
"provenance": "stack-edu-markdown-0000.json.gz:98",
"repo_name": "SamMade/BlueAndKrabby",
"revision_date": "2017-09-18T05:53:22",
"revision_id": "18abdc95d05874245e951c6f33b37e02ff11c29a",
"snapshot_id": "a229efd627a45a3b75469d2a04dc792b1be44713",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/SamMade/BlueAndKrabby/18abdc95d05874245e951c6f33b37e02ff11c29a/web/content/title-8/chapter-92F/_index.md",
"visit_date": "2021-07-01T12:51:41.871997"
} | 2.640625 | stackv2 | ---
hrs_structure:
division: '1'
volume: '2'
title: '8'
chapter: 92F
section: ''
type: chapter
tags:
- Government
menu:
hrs:
identifier: HRS0092F
parent: title8
name: Chapter 92F Uniform Information Practices Act (Modified)
weight: 55000
title: Uniform Information Practices Act (Modified)
full_title: Chapter 92F Uniform Information Practices Act (Modified)
---
**CHAPTER 92F**
**UNIFORM INFORMATION PRACTICES ACT (MODIFIED)**
Part I. General Provisions and Definitions
Section
92F-1 Short title
92F-2 Purposes; rules of construction
92F-3 General definitions
92F-4 Funding, services, and other federal assistance
Part II. Freedom of Information
92F-11 Affirmative agency disclosure responsibilities
92F-12 Disclosure required
92F-13 Government records; exceptions to general rule
92F-14 Significant privacy interest; examples
92F-15 Judicial enforcement
92F-15.3 Notice to the office of information practices
92F-15.5 Alternative method to appeal a denial of access
92F-16 Immunity from liability
92F-17 Criminal penalties
92F-18 Agency implementation
92F-19 Limitations on disclosure of government
records to other agencies
Part III. Disclosure of Personal Records
92F-21 Individual's access to own personal record
92F-21.5 Repealed
92F-22 Exemptions and limitations on individual access
92F-23 Access to personal record; initial procedure
92F-24 Right to correct personal record; initial procedure
92F-25 Correction and amendment; review procedures
92F-26 Rules
92F-27 Civil actions and remedies
92F-27.5 Alternative method to appeal a denial of access
92F-28 Access to personal records by order in judicial
or administrative proceedings; access as
authorized or required by other law
Part IV. Office of Information Practices; Duties
92F-41 Office of information practices; established
92F-42 Powers and duties of the office of information
practices
92F-43 Agency appeal of a decision by the office of
information practices
Note
Advisory opinions and guidelines interpreting this chapter are contained in the Uniform Information Practices Act Reference Manual as published by the office of information practices.
Personal information protection requirements. L Sp 2008, c 10, §§7 to 15.
Cross References
Access Hawaii Committee (management of State's internet portal), see chapter 27G.
Access/legislative information service, see chapter 21D.
Destruction of personal information records, see chapter 487R.
Information privacy and security council; personal information security, see §§487N-5 to 7.
Personal information policy and oversight responsibilities for government agencies, see §487J-5.
Personal information protection, see chapter 487J.
Privacy of consumer financial information, see §§431:3A-101 to 504.
Retail merchant club card requirements, see chapter 487D.
Security breach of personal information, see chapter 487N.
Uniform electronic transactions act, see chapter 489E.
Law Journals and Reviews
2013 Law and Administrative Rules Governing Appeal Procedures of Hawaii's Office of Information Practices. 36 UH L. Rev. 271 (2014).
Case Notes
Chapter did not require defendant to disclose development proposals. 74 H. 365, 846 P.2d 882.
Chapter applies prospectively, requiring disclosure of records maintained by state agencies regardless of when the records came into existence. 83 H. 378, 927 P.2d 386.
Chapter not a "conflicting statute on the same subject matter" as chapter 89, within the meaning of §89-19, and thus is not preempted by chapter 89 or any collective bargaining agreement negotiated under it. 83 H. 378, 927 P.2d 386.
Confidentiality provision in police union's collective bargaining agreement with city unenforceable where provision prevents police department from performing its duties under this chapter. 83 H. 378, 927 P.2d 386.
Where trial court determined that any unaccepted engineering reports were returned to the developer and there was a lack of evidence suggesting that the city planning and permitting department "maintained" any reports or copies of the reports that were unaccepted by the department, trial court properly determined that, pursuant to this chapter, the reports submitted to the department in connection with the developer's subdivision application did not constitute "government records" prior to their acceptance by the department. 119 H. 90, 194 P.3d 531.
As judicial financial disclosure statements are judicial records, created and governed by rule 15 of the rules of the supreme court, promulgated by the court pursuant to its inherent and constitutional judicial authority to regulate judges, and the authority to promulgate rules and keep records with regard to judicial qualification and discipline matters (1) fall within the scope of the court's power to adjudicate, and (2) is inherently intertwined with such adjudicative power, the rules and records are not subject to this chapter or review by the office of information practices. 121 H. 179, 215 P.3d 411.
In a suit deciding whether disclosure of county council executive session minutes was required, circuit court properly found that both chapter 92 and this chapter applied; if the meeting met an exception to the open meeting requirements put forth in chapter 92, such as an exception enumerated in §92-5, the council was not required to disclose the minutes of that meeting to the public; if the meeting did not fall under such an exception, the council was required to disclose the minutes pursuant to §92-9 and §92F-12\. 120 H. 34 (App.), 200 P.3d 403. |
2024-11-18T18:05:04.336846+00:00 | 2021-10-18T17:52:49 | cb2b9276ba6c842c34037f66c0047ebed754cf49 | 2 | {
"blob_id": "cb2b9276ba6c842c34037f66c0047ebed754cf49",
"branch_name": "refs/heads/master",
"committer_date": "2021-10-18T17:52:49",
"content_id": "3cfe5f28ae347e421e58aa798974bb5ed897477d",
"detected_licenses": [
"MIT"
],
"directory_id": "a8fb84109773bcf24026c4f09cabd04080787bf7",
"extension": "md",
"filename": "2020-06-20-ben-10-race-against-time-2007-tamil-dubbed-hd.md",
"fork_events_count": 1,
"gha_created_at": null,
"gha_event_created_at": null,
"gha_language": null,
"gha_license_id": null,
"github_id": 365491517,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 1398,
"license": "MIT",
"license_type": "permissive",
"path": "/_posts/articles/2020-06-20-ben-10-race-against-time-2007-tamil-dubbed-hd.md",
"provenance": "stack-edu-markdown-0000.json.gz:99",
"repo_name": "isaiminiya/isaiminiya.github.io",
"revision_date": "2021-10-18T17:52:49",
"revision_id": "6819bc21e9d8367d5dc3ea51deb24584c58daefe",
"snapshot_id": "de0f6447a5415272479c521d008bebf34bdf899a",
"src_encoding": "UTF-8",
"star_events_count": 0,
"url": "https://raw.githubusercontent.com/isaiminiya/isaiminiya.github.io/6819bc21e9d8367d5dc3ea51deb24584c58daefe/_posts/articles/2020-06-20-ben-10-race-against-time-2007-tamil-dubbed-hd.md",
"visit_date": "2023-08-28T19:11:25.384535"
} | 2.125 | stackv2 | ---
title: "Ben 10 Race Against Time (2007) Tamil Dubbed HD"
date: "2020-06-20"
---
## Ben 10 1 Race Against Time (2007) | Dubhoodtamil
[](https://1.bp.blogspot.com/-WtwmGjQcCCY/Xu2Be-VmBNI/AAAAAAAABc0/9WsXXWxjFesMVcIocEYDWJENMuyvCUFdwCNcBGAsYHQ/s1600/poster-780.jpg)
Movie : Ben 10 Race Against Time (2007)
Director : Alex Winter
Starring : Graham Phillips, Christien Anholt
Genres : Action, Comedy, Family
Quality : BDRip
Language : Tamil
Rating : 4.4/10
**Synopsis :** Ben Tennyson returns home from summer vacation only to discover that a rapidly aging alien criminal named, Eon, is planning to use an ancient alien artifact to destroy Bellwood.
## Click below to Download Movie
**[Download 200MB](https://oncehelp.com/ben101-200MB)**
**[Download 300MB](https://oncehelp.com/ben101-300MB)**
**Download (720p) Coming soon**
By Dubhoodtamil
Tags :
Ben 10 Race Against Time (2007) Tamil Dubbed Movie Download
Ben 10 Race Against Time (2007) TamilRockers Dubbed Movie Download
Ben 10 Race Against Time 2007 Dubbed Movie TamilGun TamilYogi Download
Ben 10 Race Against Time 2007 HD TamilDubbed Movies Download Dubhoodtamil
Ben 10 Race Against Time 2007 Hollywood Tamil Dubbed Full Movie Download
|
2024-11-18T18:05:04.583562+00:00 | 2023-08-15T06:46:56 | 1c45ab66e0ca408407b5bd65c26da7958f1a516e | 4 | {
"blob_id": "1c45ab66e0ca408407b5bd65c26da7958f1a516e",
"branch_name": "refs/heads/master",
"committer_date": "2023-08-15T06:46:56",
"content_id": "e981c44c2ae066f30e6a3ebe294c4d43ed6e2a91",
"detected_licenses": [
"MIT"
],
"directory_id": "61e760fc24431b67e8e042f186c20ac198dc1cbb",
"extension": "md",
"filename": "Multiple-Databases.md",
"fork_events_count": 85,
"gha_created_at": "2019-04-17T04:28:21",
"gha_event_created_at": "2023-08-24T13:34:50",
"gha_language": "C#",
"gha_license_id": "MIT",
"github_id": 181817060,
"is_generated": false,
"is_vendor": false,
"language": "Markdown",
"length_bytes": 2313,
"license": "MIT",
"license_type": "permissive",
"path": "/Documentation/wiki/Multiple-Databases.md",
"provenance": "stack-edu-markdown-0000.json.gz:100",
"repo_name": "dj-nitehawk/MongoDB.Entities",
"revision_date": "2023-08-15T06:46:56",
"revision_id": "aced47f36f152c6a0e13fe827f1a2bce310826dd",
"snapshot_id": "0534fc4aecccbfaec726c7ddf9afb807d37a61b3",
"src_encoding": "UTF-8",
"star_events_count": 490,
"url": "https://raw.githubusercontent.com/dj-nitehawk/MongoDB.Entities/aced47f36f152c6a0e13fe827f1a2bce310826dd/Documentation/wiki/Multiple-Databases.md",
"visit_date": "2023-08-31T20:31:26.364667"
} | 3.828125 | stackv2 | # Multiple database support
you can store and retrieve Entities in multiple databases on either a single server or multiple servers. the only requirement is to have unique names for each database. the following example demonstrates how to use multiple databases.
### Usage example:
use the `DB.DatabaseFor<T>()` method to specify which database you want the Entities of a given type to be stored in. it is only neccessary to do that for the entities you want to store in a non-default database. the default database is the very first database your application initializes. all entities by default are stored in the default database unless specified otherwise using `DatabaseFor`.
as such, the `Book` entities will be stored in the "BookShop" database and the `Picture` entities are stored in the "BookShopFILES" database considering the following code.
```csharp
await DB.InitAsync("BookShop");
await DB.InitAsync("BookShopFILES");
DB.DatabaseFor<Picture>("BookShopFILES");
var book = new Book { Title = "Power Of Now" };
await book.SaveAsync();
//alternative:
//// await DB.SaveAsync(book);
var pic = new Picture
{
BookID = book.ID,
Name = "Power Of Now Cover Photo"
};
await pic.SaveAsync();
//alternative:
//// await DB.SaveAsync(pic);
await DB.Update<Picture>()
.Match(p => p.ID == pic.ID)
.Modify(p => p.Name, "Updated Cover Photo")
.ExecuteAsync();
var result = await DB.Find<Picture>().OneAsync(pic.ID);
```
> [!note]
> an entity type is tied to a specific database by calling the **DatabaseFor** method with the database name on startup. that entity type will always be stored in and retrieved from that specific database only. it is not possible to save a single entity type in multiple databases.
if you prefer to keep your database specifications inside the entity classes themselves, you could even call `DatabaseFor` in the static constructor like so:
```csharp
public class Picture : Entity
{
static Picture() => DB.DatabaseFor<Picture>("BookShopFILES");
}
```
### Limitations
- cross-database relationships with `Many<T>` is not supported.
- no cross-database joins/ look-ups as the driver doesn't support it.
- storing a single entity type in multiple datbases is not supported. |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.