From a65b1ff578dffe954fbfebcb51b8870d66805a24 Mon Sep 17 00:00:00 2001 From: Patrick Jentsch
- nopaque was designed from the ground up to be modular. This modularity - means that the offered workflow provides variable entry and exit points, - so that different starting points and goals can be flexibly addressed. - Each of these modules are implemented in a self-contained service, each of - which represents a step in the workflow. The services are coordinated in - such a way that they can be used consecutively. The order can either be - taken from the listing of the services in the left sidebar or from the - roadmap (accessible via the pink compass in the upper right corner). All - services are versioned, so the data generated with nopaque is always - reproducible. -
-- The File Setup Service bundles image data, such as scans and photos, - together in a handy PDF file. To use this service, use the job form to - select the images to be bundled, choose the desired service version, and - specify a title and description. Please note that the service sorts the - images into the resulting PDF file based on the file names. So naming the - images correctly is of great importance. It has proven to be a good practice - to name the files according to the following scheme: - page-01.png, page-02.jpg, page-03.tiff, etc. In general, you can assume - that the images will be sorted in the order in which the file explorer of - your operating system lists them when you view the files in a folder - sorted in ascending order by file name. -
- -Coming soon...
- -Coming soon...
- -Coming soon...
- -- With the corpus analysis service, it is possible to create a text corpus - and then explore it in an analysis session. The analysis session is realized - on the server side by the Open Corpus Workbench software, which enables - efficient and complex searches with the help of the CQP Query Language. -
diff --git a/app/templates/_base/_modals/_manual/07_a_closer_look_at_the_corpus_analysis.html.j2 b/app/templates/_base/_modals/_manual/07_a_closer_look_at_the_corpus_analysis.html.j2 deleted file mode 100644 index db9b6e5f..00000000 --- a/app/templates/_base/_modals/_manual/07_a_closer_look_at_the_corpus_analysis.html.j2 +++ /dev/null @@ -1,47 +0,0 @@ -- To create a corpus, you - can use the "New Corpus" button, which can be found on both the Corpus - Analysis Service page and the Dashboard below the corpus list. Fill in the input - mask to Create a corpus. After you have completed the input mask, you will - be automatically taken to the corpus overview page (which can be called up - again via the corpus lists) of your new and accordingly still empty corpus. -
-- Now you can add texts in vrt format (results of the NLP service) to your new - corpus. To do this, use the "Add Corpus File" button and fill in the form - that appears. You will get the possibility to add metadata to each text. - After you have added all the desired texts to the corpus, the corpus must be - prepared for the analysis, this process can be initiated by clicking on the - "Build" button. On the corpus overview page you can always see information - about the current status of the corpus in the upper right corner. After the - build process the status should be "built". -
-- After you have created and built a corpus, it can be analyzed. To do this, - use the button labeled Analyze. The corpus analysis currently offers two - modules, the Reader and the Concordance module. The reader module can be - used to read your tokenized corpus in different ways. You can select a token - representation option, it determines the property of a token to be shown. - You can for example read your text completly lemmatized. You can also change - the way of how a token is displayed, by using the text style switch. The - concordance module offers some more options regarding the context size of - search results. If the context does not provide enough information you can - hop into the reader module by using the magnifier icon next to a match. -
diff --git a/app/templates/_base/_modals/_manual/09_query_builder.html.j2 b/app/templates/_base/_modals/_manual/09_query_builder.html.j2 deleted file mode 100644 index 31f46f52..00000000 --- a/app/templates/_base/_modals/_manual/09_query_builder.html.j2 +++ /dev/null @@ -1,211 +0,0 @@ -The query builder can be accessed via "My Corpora" or "Corpus Analysis" in the sidebar options. -Select the desired corpus and click on the "Analyze" and then "Concordance" -buttons to open the query builder.
-The query builder uses the Corpus Query Language (CQL) to help you make a query for analyzing your texts. -In this way, it is possible to filter out various types of text parameters, for -example, a specific word, a lemma, or you can set part-of-speech -tags (pos) that indicate the type of word you are looking for (a noun, an -adjective, etc.). In addition, you can also search for structural attributes, -or specify your query for a token (word, lemma, pos) via entity typing. And of -course, the different text parameters can be combined.
-Tokens and structural attributes can be added by clicking on the "+" button
-(the "input marker") in the input field or the labeled buttons below it. Elements
-added are shown as chips. These can be reorganized using drag and drop. The input
-marker can also be moved in this way. Its position shows where new elements will be added.
-A "translation" of your query into Corpus Query Language (CQL) is shown below.
Advanced users can make direct use of the Corpus Query Language (CQL) by switching to "expert mode" via the toggle button.
-The entire input field can be cleared using the red trash icon on the right.
-If you are only looking for a specific token, you can click on the left - button and select the type of token you are looking for from the drop-down menu. - By default "Word" is selected.
-If you want to search for a specific word or lemma and the respective - category is selected in the drop-down menu, you can type in the word or lemma - of your choice in the input field. You can confirm your entry by clicking the - Submit button on the right. You can also use the options below to modify your - token request before pressing the submit button. These options are explained - further here.
- -You can choose between the options "english pos", "german pos" and - "simple_pos" to search for different parts-of-speech. You can find an overview - of all tags under the "Tagsets" tab.
- -Here you can search for a token with unspecified attributes (also called wildcard token). This - selection should never stand alone and should always be extended with an incidence modifier or stand in a - larger query, because otherwise all possible tokens would be searched for and - the program would crash.
-You have the possibility to extend or specify the token you are searching for with - certain factors. For this the query builder offers some fixed options. You can - find more information about the options in the Corpus Query Language Tutorial.
-A wildcard character replaces any character and is represented in the form of a dot.
-With an option group you can search for different variants of a token. The - variants are not limited, so you can manually enter more options in the same - format. "Option1" and "option2" must be replaced accordingly.
- -With the Incidence Modifiers you can determine the occurrence of single
- tokens. For example you can use "?" to indicate that the token occurs either
- not at all or once:
- [word = "is"] [word="it"] [word="your"] [word="litte"]? [word = "dog"]
- Here the word "little" should occur either once or not at all. With
- [word="dogs?"] the search is for "dog "or "dogs".
With the check mark at Ignore Case the upper and lower case is ignored. - This is marked with a "%c". By default (if not checked) it is case sensitive.
-"Or" ( | ) and "and" ( & ) are conditions you can put on a token. With "or"
- one of the two conditions must be fulfilled, with "and" both conditions must be
- fulfilled. For example, the word must be called "will" AND be a verb, only then
- it will be displayed. Note that "and" is not responsible for lining up tokens in
- this case. For this you can simply string them together:
- [word="I"] [word="will" & simple_pos="VERB"] [word="go"].
Tokens that have already been added can also be modified by clicking on the corresponding - pen icon. Click on the "ignore case" box, for example, and the query builder will - not differentiate between upper- and lower- case letters for that respective token. - New conditions added apply to the most recent token information.
-You can use structural attributes to search specifically for structures in - the text or to further narrow down your previous search query.
-With entities, i.e. units of meaning, you can search for text sections that
- contain more specific information, for example, persons, dates, or events. The
- codes for these categories can be selected using the drop-down menus. You can find an explanation of
- these abbreviations under the tab "Tagsets".
- You can also search for unspecified entities by selecting "Add entity of any type".
The entity type can be changed by clicking on the pen symbol on the chip. When - the Entity attribute is added, the input marker will automatically be - moved between the entity chips. Use drag-and-drop as needed to continue your query - at a different position.
- - -The meta data function is being worked on and cannot currently be used!
-
- With the meta data you can annotate your text and add specific conditions.
- You can select a category on the left and enter your desired value on the right.
- The selected metadata will apply to your entire request and will be added at the end.
You have several options to edit your query after adding it to the preview.
-You can edit your query chips by clicking on the pen icon.
- -You can delete the added elements from the query by clicking the X behind the respective content.
- -You can drag and drop elements to customize your query.
- -With the incidence modifier option, you can specify the amount of - times a token should appear in your query. This is particularly relevant for empty - tokens (tokens with unspecified attributes). Click on a token (blue chip) and - select the desired option from the list to add an incidence modifier. To - close the list without adding anything, click on the token again.
- -To work with the plain Corpus Query Language instead of using the Query Builder, click on the "expert mode" - switch. Your query can be entered into the input field. All elements previously added will be carried over - into expert mode. Click on the switch again to switch back to the Query Builder if desired. All recognized elements - will be parsed into chips; those not recognized will be deleted from the query.
- - - -nopaque is a web-based digital working environment. It implements a workflow based on the research process in the humanities and supports its diff --git a/app/templates/_base/_modals/_manual/02_registration_and_log_in.html.j2 b/app/templates/main/_manual/02_registration_and_log_in.html.j2 similarity index 92% rename from app/templates/_base/_modals/_manual/02_registration_and_log_in.html.j2 rename to app/templates/main/_manual/02_registration_and_log_in.html.j2 index 5f05c543..b50d1546 100644 --- a/app/templates/_base/_modals/_manual/02_registration_and_log_in.html.j2 +++ b/app/templates/main/_manual/02_registration_and_log_in.html.j2 @@ -1,8 +1,10 @@ -
Before you can start using the web platform, you need to create a user diff --git a/app/templates/_base/_modals/_manual/03_dashboard.html.j2 b/app/templates/main/_manual/03_dashboard.html.j2 similarity index 96% rename from app/templates/_base/_modals/_manual/03_dashboard.html.j2 rename to app/templates/main/_manual/03_dashboard.html.j2 index 51d772a3..e53e8877 100644 --- a/app/templates/_base/_modals/_manual/03_dashboard.html.j2 +++ b/app/templates/main/_manual/03_dashboard.html.j2 @@ -1,8 +1,10 @@ -
The dashboard provides a central overview of all resources assigned to the @@ -12,7 +14,9 @@ the search field displayed above them.
+ nopaque was designed from the ground up to be modular. This modularity + means that the offered workflow provides variable entry and exit points, + so that different starting points and goals can be flexibly addressed. + Each of these modules are implemented in a self-contained service, each of + which represents a step in the workflow. The services are coordinated in + such a way that they can be used consecutively. The order can either be + taken from the listing of the services in the left sidebar or from the + roadmap (accessible via the pink compass in the upper right corner). All + services are versioned, so the data generated with nopaque is always + reproducible. +
++ The File Setup Service bundles image data, such as scans and photos, + together in a handy PDF file. To use this service, use the job form to + select the images to be bundled, choose the desired service version, and + specify a title and description. Please note that the service sorts the + images into the resulting PDF file based on the file names. So naming the + images correctly is of great importance. It has proven to be a good practice + to name the files according to the following scheme: + page-01.png, page-02.jpg, page-03.tiff, etc. In general, you can assume + that the images will be sorted in the order in which the file explorer of + your operating system lists them when you view the files in a folder + sorted in ascending order by file name. +
+ +Coming soon...
+ +Coming soon...
+ +Coming soon...
+ ++ With the corpus analysis service, it is possible to create a text corpus + and then explore it in an analysis session. The analysis session is realized + on the server side by the Open Corpus Workbench software, which enables + efficient and complex searches with the help of the CQP Query Language. +
+ ++ To create a corpus, you + can use the "New Corpus" button, which can be found on both the Corpus + Analysis Service page and the Dashboard below the corpus list. Fill in the input + mask to Create a corpus. After you have completed the input mask, you will + be automatically taken to the corpus overview page (which can be called up + again via the corpus lists) of your new and accordingly still empty corpus. +
++ Now you can add texts in vrt format (results of the NLP service) to your new + corpus. To do this, use the "Add Corpus File" button and fill in the form + that appears. You will get the possibility to add metadata to each text. + After you have added all the desired texts to the corpus, the corpus must be + prepared for the analysis, this process can be initiated by clicking on the + "Build" button. On the corpus overview page you can always see information + about the current status of the corpus in the upper right corner. After the + build process the status should be "built". +
++ After you have created and built a corpus, it can be analyzed. To do this, + use the button labeled Analyze. The corpus analysis currently offers two + modules, the Reader and the Concordance module. The reader module can be + used to read your tokenized corpus in different ways. You can select a token + representation option, it determines the property of a token to be shown. + You can for example read your text completly lemmatized. You can also change + the way of how a token is displayed, by using the text style switch. The + concordance module offers some more options regarding the context size of + search results. If the context does not provide enough information you can + hop into the reader module by using the magnifier icon next to a match. +
diff --git a/app/templates/_base/_modals/_manual/08_cqp_query_language.html.j2 b/app/templates/main/_manual/08_cqp_query_language.html.j2 similarity index 97% rename from app/templates/_base/_modals/_manual/08_cqp_query_language.html.j2 rename to app/templates/main/_manual/08_cqp_query_language.html.j2 index 9069d495..0af61dd6 100644 --- a/app/templates/_base/_modals/_manual/08_cqp_query_language.html.j2 +++ b/app/templates/main/_manual/08_cqp_query_language.html.j2 @@ -1,4 +1,4 @@ -Within the Corpus Query Language, a distinction is made between two types of annotations: positional attributes and structural attributes. Positional attributes refer to a token, e.g. the word "book" is assigned the part-of-speech tag "NN", the lemma "book" and the simplified part-of-speech tag "NOUN" within the token structure. Structural attributes refer to text structure-giving elements such as sentence and entity markup. For example, the markup of a sentence is represented in the background as follows:
@@ -13,7 +13,7 @@
-Before you can start searching for positional attributes (also called tokens), it is necessary to know what properties they contain.
Token with no condition on any property (also called wildcard token)
@@ -118,7 +118,7 @@
^ ^ the braces indicate the start and end of an option group
nopaque provides several structural attributes for query. A distinction is made between attributes with and without value.
<ent> [] </ent>; A one token long entity of any type
<ent_type="PERSON"> [] </ent_type>; A one token long entity of type PERSON
<ent_type="PERSON"> []* </ent_type>; Entity of any length of type PERSON
diff --git a/app/templates/main/_manual/09_query_builder.html.j2 b/app/templates/main/_manual/09_query_builder.html.j2
new file mode 100644
index 00000000..8d413f0e
--- /dev/null
+++ b/app/templates/main/_manual/09_query_builder.html.j2
@@ -0,0 +1,251 @@
++ The query builder can be accessed via "My Corpora" or "Corpus Analysis" in the sidebar options. + Select the desired corpus and click on the "Analyze" and then "Concordance" + buttons to open the query builder. +
+ ++ The query builder uses the Corpus Query Language (CQL) to help you make a query for analyzing your texts. + In this way, it is possible to filter out various types of text parameters, for + example, a specific word, a lemma, or you can set part-of-speech + tags (pos) that indicate the type of word you are looking for (a noun, an + adjective, etc.). In addition, you can also search for structural attributes, + or specify your query for a token (word, lemma, pos) via entity typing. And of + course, the different text parameters can be combined. +
+ +
+ Tokens and structural attributes can be added by clicking on the "+" button
+ (the "input marker") in the input field or the labeled buttons below it. Elements
+ added are shown as chips. These can be reorganized using drag and drop. The input
+ marker can also be moved in this way. Its position shows where new elements will be added.
+ A "translation" of your query into Corpus Query Language (CQL) is shown below.
+
Advanced users can make direct use of the Corpus Query Language (CQL) by switching to "expert mode" via the toggle button.
+The entire input field can be cleared using the red trash icon on the right.
+ ++ If you are only looking for a specific token, you can click on the left + button and select the type of token you are looking for from the drop-down menu. + By default "Word" is selected. +
++ If you want to search for a specific word or lemma and the respective + category is selected in the drop-down menu, you can type in the word or lemma + of your choice in the input field. You can confirm your entry by clicking the + Submit button on the right. You can also use the options below to modify your + token request before pressing the submit button. These options are explained + further here. +
+ ++ You can choose between the options "english pos", "german pos" and + "simple_pos" to search for different parts-of-speech. You can find an overview + of all tags under the "Tagsets" tab. +
+ ++ Here you can search for a token with unspecified attributes (also called wildcard token). This + selection should never stand alone and should always be extended with an incidence modifier or stand in a + larger query, because otherwise all possible tokens would be searched for and + the program would crash. +
++ You have the possibility to extend or specify the token you are searching for with + certain factors. For this the query builder offers some fixed options. You can + find more information about the options in the Corpus Query Language Tutorial. +
+A wildcard character replaces any character and is represented in the form of a dot.
++ With an option group you can search for different variants of a token. The + variants are not limited, so you can manually enter more options in the same + format. "Option1" and "option2" must be replaced accordingly. +
+ +
+ With the Incidence Modifiers you can determine the occurrence of single
+ tokens. For example you can use "?" to indicate that the token occurs either
+ not at all or once:
+ [word = "is"] [word="it"] [word="your"] [word="litte"]? [word = "dog"]
+ Here the word "little" should occur either once or not at all. With
+ [word="dogs?"] the search is for "dog "or "dogs".
+
+ With the check mark at Ignore Case the upper and lower case is ignored. + This is marked with a "%c". By default (if not checked) it is case sensitive. +
+
+ "Or" ( | ) and "and" ( & ) are conditions you can put on a token. With "or"
+ one of the two conditions must be fulfilled, with "and" both conditions must be
+ fulfilled. For example, the word must be called "will" AND be a verb, only then
+ it will be displayed. Note that "and" is not responsible for lining up tokens in
+ this case. For this you can simply string them together:
+ [word="I"] [word="will" & simple_pos="VERB"] [word="go"].
+
+ Tokens that have already been added can also be modified by clicking on the corresponding + pen icon. Click on the "ignore case" box, for example, and the query builder will + not differentiate between upper- and lower- case letters for that respective token. + New conditions added apply to the most recent token information. +
++ You can use structural attributes to search specifically for structures in + the text or to further narrow down your previous search query. +
+
+ With "Sentence" (<s></s>) you can search for sentences within your text.
+ This search can of course be specified if you search for particular tokens or
+ entities between the sentence tags (<s></s>). For example, you can search for
+ sentences that contain only a noun, verb, and adjective.
+ Click on Sentence to add the sentence chips: Sentence Start
+ and Sentence End.
+ These mark where the sentence starts and ends. Use drag-and-drop to place them accordingly. When
+ the Sentence attribute is added, the input marker will automatically be
+ moved between the sentence chips. Use drag-and-drop as needed to continue your query
+ at a different position.
+
+ With entities, i.e. units of meaning, you can search for text sections that
+ contain more specific information, for example, persons, dates, or events. The
+ codes for these categories can be selected using the drop-down menus. You can find an explanation of
+ these abbreviations under the tab "Tagsets".
+ You can also search for unspecified entities by selecting "Add entity of any type".
+
+ Click on the Entity button to add the entity chips Entity Type= + and Entity End. +
+ ++ The entity type can be changed by clicking on the pen symbol on the chip. When + the Entity attribute is added, the input marker will automatically be + moved between the entity chips. Use drag-and-drop as needed to continue your query + at a different position. +
+ +
+ The meta data function is being worked on and cannot currently be used!
+ With the meta data you can annotate your text and add specific conditions.
+ You can select a category on the left and enter your desired value on the right.
+ The selected metadata will apply to your entire request and will be added at the end.
+
You have several options to edit your query after adding it to the preview.
+You can edit your query chips by clicking on the pen icon.
+ +You can delete the added elements from the query by clicking the X behind the respective content.
+ +You can drag and drop elements to customize your query.
+ ++ With the incidence modifier option, you can specify the amount of + times a token should appear in your query. This is particularly relevant for empty + tokens (tokens with unspecified attributes). Click on a token (blue chip) and + select the desired option from the list to add an incidence modifier. To + close the list without adding anything, click on the token again. +
+ ++ To work with the plain Corpus Query Language instead of using the Query Builder, click on the "expert mode" + switch. Your query can be entered into the input field. All elements previously added will be carried over + into expert mode. Click on the switch again to switch back to the Query Builder if desired. All recognized elements + will be parsed into chips; those not recognized will be deleted from the query. +
+ +