Library

Your email was sent successfully. Check your inbox.

An error occurred while sending the email. Please try again.

Proceed reservation?

Export
Filter
  • 2005-2009  (17)
  • 1985-1989
  • 1870-1879
  • 2009  (8)
  • 2008  (9)
  • ddc:000  (12)
  • ddc:620  (5)
Source
Years
  • 2005-2009  (17)
  • 1985-1989
  • 1870-1879
Year
Keywords
Language
  • 1
    Publication Date: 2016-06-09
    Description: In this paper we revisit the a priori turbulent flame speed tabulation (TFST) technique for a given parameter space within the region of flamelet combustion-regimes. It can be used as a subgrid-scale (SGS) model in Large Eddy Simulation (LES). In a first step, stationary laminar flamelets are computed and stored over the progress variable following the ideas of flamelet generated manifolds (FGM). In a second step, the incompressible one-dimensional Navier-Stokes equations supplemented by the equation for the progress variable are solved on a grid that resolves all turbulent scales. Additionally, turbulent transport is implemented via the linear eddy model (LEM). The turbulent flame structures are solved until a statistically stationary mean value of the turbulent flame speed has been reached. The results are stored in a table that could be used by large scale premixed combustion models, e.g. front tracking schemes. First results are compared quantitatively with direct numerical simulations (DNS) taken from literature. Then it is illustrated in one example how the new method could help to fix constants in algebraic models for turbulent flame speeds. Further it is shown how the technique can be extended to incorporate turbulent strain effects. Finally we investigate the effect of the use of detailed and tabulated chemistry under unsteady conditions.
    Keywords: ddc:620
    Language: English
    Type: reportzib , doc-type:preprint
    Format: application/pdf
    Library Location Call Number Volume/Issue/Year Availability
    BibTip Others were also interested in ...
  • 2
    Publication Date: 2020-12-15
    Description: The Steiner connectivity problem is a generalization of the Steiner tree problem. It consists in finding a minimum cost set of simple paths to connect a subset of nodes in an undirected graph. We show that polyhedral and algorithmic results on the Steiner tree problem carry over to the Steiner connectivity problem, namely, the Steiner cut and the Steiner partition inequalities, as well as the associated polynomial time separation algorithms, can be generalized. Similar to the Steiner tree case, a directed formulation, which is stronger than the natural undirected one, plays a central role.
    Keywords: ddc:000
    Language: English
    Type: reportzib , doc-type:preprint
    Format: application/pdf
    Format: application/postscript
    Format: application/pdf
    Library Location Call Number Volume/Issue/Year Availability
    BibTip Others were also interested in ...
  • 3
    Publication Date: 2020-12-11
    Description: We give the basic definitions and some theoretical results about hyperdeterminants, introduced by A.~Cayley in 1845. We prove integrability (understood as $4d$-consistency) of a nonlinear difference equation defined by the $2 \times 2 \times 2$ - hyperdeterminant. This result gives rise to the following hypothesis: the difference equations defined by hyperdeterminants of any size are integrable. We show that this hypothesis already fails in the case of the $2\times 2\times 2\times 2$ - hyperdeterminant.
    Keywords: ddc:000
    Language: English
    Type: reportzib , doc-type:preprint
    Format: application/pdf
    Format: application/postscript
    Library Location Call Number Volume/Issue/Year Availability
    BibTip Others were also interested in ...
  • 4
    Publication Date: 2016-06-09
    Description: A world-wide used program for the simulation of fire-induced flows is the Fire Dynamics Simulator (FDS) which originally was developed for a purely serial execution on single-processor computing systems. Due to steadily increasing problem sizes and accuracy requirements as well as restrictions in storage capacity and computing power on single-processor systems, the efficient simulation of the considered fire scenarios can only be achieved on modern high-performance systems based on multi-processor architectures. The transition to those systems requires the elaborate parallelization of the underlying numerical methods which must guarantee the same result for a given problem as the corresponding serial execution. Unfortunately, one fundamental serial serial solver of FDS, the pressure solver, only possesses a low degree of inherent parallelizm. Its current parallelization may cause additional numerical errors, casually leading to significant losses of accuracy or even numerical instabilities. In order to ensure that the parallelization errors are limited by the leading error of the numerical scheme such that second order convergence for the whole method can be acchieved, optimized parallelization concepts must be designed. With respect to these considerations this articles gives an overview of the current parallel pressure solver as well as the problems related to it and presents an alternative method, SCARC, to overcome the existing complicacies. Part I explains the theory, concept and implementation of this new strategy, whereas Part II describes a series of validation and verification tests to proof its correctness.
    Keywords: ddc:620
    Language: English
    Type: reportzib , doc-type:preprint
    Format: application/pdf
    Library Location Call Number Volume/Issue/Year Availability
    BibTip Others were also interested in ...
  • 5
    Publication Date: 2016-06-09
    Description: Because CFD programs, like FDS, generally consist of a large number of different components representing the variety of participating numerical algorithms and chemical / physical processes, it is nearly impossible to verify such codes in their entirety, for example with comparisons of fire tests. Instead, a careful verification and validation with respect to the underlying mathematical conditions and applied numerical schemes is indispensable. In particular, error cancelations between single program components can only be detected by such detailed component-level tests. In part I of this article series a conceptual deficiency of the FDS program package with regard to multi-mesh computations was illustrated and an alternative domain decomposition strategy FDS-ScaRC was introduced. In this second part we will present the structure of a comprehensive test concept and the needs for a more mathematically and numerically orientated test procedure that is much more suited for a reliable evaluation than only a simple visual comparison of the numerical results with experimental fire tests. After a general introduction of our test concept we will demonstrate the high potential of the new FDS-\scarc{} technique compared to the FDS-FFT technique which is used in the FDS program package as yet. Based on this concept, we will present a comprehensive set of analytical and numerical test results.
    Keywords: ddc:620
    Language: English
    Type: reportzib , doc-type:preprint
    Format: application/pdf
    Library Location Call Number Volume/Issue/Year Availability
    BibTip Others were also interested in ...
  • 6
    Publication Date: 2016-06-09
    Description: Starting from the conservation laws for mass, momentum and energy together with a three species, bulk microphysic model, a model for the interaction of internal gravity waves and deep convective hot towers is derived by using multiscale asymptotic techniques. From the resulting leading order equations, a closed model is obtained by applying weighted averages to the smallscale hot towers without requiring further closure approximations. The resulting model is an extension of the linear, anelastic equations, into which moisture enters as the area fraction of saturated regions on the microscale with two way coupling between the large and small scale. Moisture reduces the effective stability in the model and defines a potential temperature sourceterm related to the net effect of latent heat release or consumption by microscale up- and downdrafts. The dispersion relation and group velocity of the system is analyzed and moisture is found to have several effects: It reduces energy transport by waves, increases the vertical wavenumber but decreases the slope at which wave packets travel and it introduces a lower horizontal cutoff wavenumber, below which modes turn into evanescent. Further, moisture can cause critical layers. Numerical examples for steadystate and timedependent mountain waves are shown and the effects of moisture on these waves are investigated.
    Keywords: ddc:000
    Language: English
    Type: reportzib , doc-type:preprint
    Format: application/pdf
    Format: application/postscript
    Library Location Call Number Volume/Issue/Year Availability
    BibTip Others were also interested in ...
  • 7
    Publication Date: 2020-08-05
    Description: In this paper we investigate the performance of several out-of-the box solvers for mixed-integer quadratically constrained programmes (MIQCPs) on an open pit mine production scheduling problem with mixing constraints. We compare the solvers BARON, Couenne, SBB, and SCIP to a problem-specific algorithm on two different MIQCP formulations. The computational results presented show that general-purpose solvers with no particular knowledge of problem structure are able to nearly match the performance of a hand-crafted algorithm.
    Keywords: ddc:000
    Language: English
    Type: reportzib , doc-type:preprint
    Format: application/pdf
    Format: application/postscript
    Library Location Call Number Volume/Issue/Year Availability
    BibTip Others were also interested in ...
  • 8
    Publication Date: 2020-12-11
    Description: Eines der Hauptanliegen bei der Initiierung des Kooperativen Bibliotheksverbundes Berlin-Brandenburg (KOBV) war die Schaffung einer modernen, leistungsfähigen Informationsinfrastruktur unter Einbeziehung möglichst aller Bibliotheken der in ihrer Dichte und Vielfalt in Deutschland einmaligen Bibliothekslandschaft. Sie wurde beim Aufbau des KOBV durch die Wahl des Verbundmodelles mit seinen offenen, integrativen Strukturen befördert. Die Ziele und Strategien des KOBV leiten sich aus dem Grundauftrag der Bibliotheken her, ihre NutzerInnen optimal mit Medien und Informationen zu versorgen. Mit dem KOBV haben die Bibliotheken eine gemeinsame Plattform und ein regionales Netzwerk geschaffen, mit deren Hilfe sie die Informationsinfrastruktur in Berlin und Brandenburg zusammen aufgebaut und grundlegend verbessert haben. Nach rund elf Jahren ist der KOBV in eine neue Phase eingetreten, indem er – unter Wahrung seiner institutionellen Eigenständigkeit – eine Strategische Allianz mit dem Bibliotheksverbund Bayern (BVB) eingegangen ist und die bislang auf die Region bezogene Kooperation der Bibliotheken überregional ausgeweitet hat. Wenn auch die Universalbibliotheken des KOBV in dieser Zusammenarbeit einen Kurswechsel zur zentralen Katalogisierung vollziehen, bedeutet dies für den KOBV insgesamt keinen Paradigmenwechsel: Er bleibt ein dezentraler Verbund und wird sein flexibles technisches Mischkonzept von verteilter und zentraler Datenhaltung fortführen. Dieses lässt den Bibliotheken die Freiheit zu entscheiden, ob sie zentral oder dezentral katalogisieren möchten. Die KOBV-Zentrale, die im KOBV keine Aufgaben in der Katalogisierung wahrnimmt, wird wie bislang allen Mitgliedsbibliotheken die ASP-Dienste und weiteren Angebote gleichermaßen zur Verfügung stellen und die Dienstleistungen für NutzerInnen und Bibliotheken weiter ausbauen. Mit dem vorhandenen Know How, der eingesetzten modernen Technologie und der strategischen Entscheidung, Ressourcen und Kompetenzen mit dem BVB zu bündeln und durch verbundübergreifende arbeitsteilige Verfahren Synergien zu gewinnen, ist der KOBV gut aufgestellt, um die Herausforderungen der kommenden Jahre anzugehen und das Leistungsspektrum der KOBV-Bibliotheken durch die Entwicklung neuer Dienste auszuweiten.
    Keywords: ddc:000
    Language: German
    Type: reportzib , doc-type:preprint
    Format: application/pdf
    Library Location Call Number Volume/Issue/Year Availability
    BibTip Others were also interested in ...
  • 9
    Publication Date: 2020-08-05
    Description: We consider an auction of slots to run trains through a railway network. In contrast to the classical setting for combinatorial auctions, there is not only competition for slots, but slots can mutually exclude each other, such that general conflict constraints on bids arise. This turns the winner determination problem associated with such an auction into a complex combinatorial optimization problem. It also raises a number of auction design questions, in particular, on incentive compatibilty. We propose a single-shot second price auction for railway slots, the Vickrey Track Auction (VTA). We show that this auction is incentive compatible, i.e., rational bidders are always motivated to bid their true valuation, and that it produces efficient allocations, even in the presence of constraints on allocations. These properties are, however, lost when rules on the submission of bids such as, e.g., lowest bids, are imposed. Our results carry over to generalized" Vickrey auctions with combinatorial constraints.
    Keywords: ddc:000
    Language: English
    Type: reportzib , doc-type:preprint
    Format: application/pdf
    Format: application/postscript
    Library Location Call Number Volume/Issue/Year Availability
    BibTip Others were also interested in ...
  • 10
    Publication Date: 2020-08-05
    Description: Technical restrictions and challenging details let railway traffic become one of the most complex transportation systems. Routing trains in a conflict-free way through a track network is one of the basic scheduling problems for any railway company. This article focuses on a robust extension of this problem, also known as train timetabling problem (TTP), which consists in finding a schedule, a conflict free set of train routes, of maximum value for a given railway network. However, timetables are not only required to be profitable. Railway companies are also interested in reliable and robust solutions. Intuitively, we expect a more robust track allocation to be one where disruptions arising from delays are less likely to be propagated causing delays of subsequent trains. This trade-off between an efficient use of railway infrastructure and the prospects of recovery leads us to a bi-criteria optimization approach. On the one hand we want to maximize the profit of a schedule, that is more or less to maximize the number of feasible routed trains. On the other hand if two trains are scheduled as tight as possible after each other it is clear that a delay of the first one always affects the subsequent train. We present extensions of the integer programming formulation in [BorndoerferSchlechte2007] for solving (TTP). These models can incorporate both aspects, because of the additional track configuration variables. We discuss how these variables can directly be used to measure a certain type of robustness of a timetable. For these models which can be solved by column generation techniques, we propose so-called scalarization techniques, see [Ehrgott2005], to determine efficient solutions. Here, an efficient solution is one which does not allow any improvement in profit and robustness at the same time. We prove that the LP-relaxation of the (TTP) including an additional $\epsilon$-constraint remains solvable in polynomial time. Finally, we present some preliminary results on macroscopic real-world data of a part of the German long distance railway network.
    Keywords: ddc:000
    Language: English
    Type: reportzib , doc-type:preprint
    Format: application/pdf
    Format: application/pdf
    Format: application/postscript
    Format: application/postscript
    Library Location Call Number Volume/Issue/Year Availability
    BibTip Others were also interested in ...
  • 11
    Publication Date: 2016-06-09
    Description: In this paper we propose a technique for a priori turbulent flame speed tabulation (TFST) for a given parameter space in standard combustion-regime diagrams. It can be used as a subgrid-scale (SGS) model in Large Eddy Simulation (LES). In a first step, stationary laminar flamelets are computed and stored over the progress variable following the ideas of flamelet generated manifolds (FGM). In a second step, the incompressible one-dimensional Navier-Stokes equations supplemented by the equation for the progress variable are solved on a grid that resolves all turbulent scales. Additionally, turbulent transport is implemented via the linear eddy model (LEM). The turbulent flame structures are solved until a statistically stationary mean value of the turbulent flame speed has been reached. The results are stored in a table that could be used by large scale premixed combustion models, e.g. front tracking schemes. Results are compared to an algebraic model and to direct numerical simulations (DNS).
    Keywords: ddc:620
    Language: English
    Type: reportzib , doc-type:preprint
    Format: application/pdf
    Library Location Call Number Volume/Issue/Year Availability
    BibTip Others were also interested in ...
  • 12
    Publication Date: 2016-06-09
    Description: The influence of thermal stratification on autoignition at constant volume and high pressure is investigated under turbulent conditions using the one-dimensional Linear-Eddy Model (LEM) and detailed hydrogen/air chemistry. Results are presented for the influence of initial temperature inhomogeneities on the heat release rate and the relative importance of diffusion and chemical reactions. The predicted heat release rates are compared with heat release rates of Chen et al. and Hawkes et al. obtained by two-dimensional Direct Numerical Simulations (DNS). Using the definition of Chen et al. for the displacement speed of the H2 mass fraction tracked at the location of maximum heat release, and a comparison of budget terms, different combustion modes including ignition front propagation and deflagration waves are identified and the results are compared to the DNS data. The LEM approach shows qualitatively and quantitatively reasonable agreement with the DNS data over the whole range of investigated temperature fluctuations. The results presented in this work suggest that LEM is a potential candidate as a sub-model for CFD calculations of HCCI engines.
    Keywords: ddc:620
    Language: English
    Type: reportzib , doc-type:preprint
    Format: application/pdf
    Library Location Call Number Volume/Issue/Year Availability
    BibTip Others were also interested in ...
  • 13
    Publication Date: 2016-06-30
    Description: The latest machine generation installed at supercomputer centres in Germany offers a peak performance in the tens of Tflop/s range. We study performance and scaling of our quantum chromodynamics simulation programme BQCD that we obtained on two of these machines, an IBM Blue Gene/L and an SGI Altix 4700. We compare the performance of Fortran/MPI code with assembler code. The latter allows to exploit concurrency at more levels, in particular in overlapping communication and computation as well as prefetching data from main memory.
    Keywords: ddc:000
    Language: English
    Type: reportzib , doc-type:preprint
    Format: application/pdf
    Format: application/postscript
    Library Location Call Number Volume/Issue/Year Availability
    BibTip Others were also interested in ...
  • 14
    Publication Date: 2016-06-09
    Description: We study the optimal control of a maximum-norm objective functional subject to an elliptic-type PDE and pointwise state constraints. The problem is transformed into a problem where the non-differentiable L^{\infty}-norm in the functional will be replaced by a scalar variable and additional state constraints. This problem is solved by barrier methods. We will show the existence and convergence of the central path for a class of barrier functions. Numerical experiments complete the presentation.
    Keywords: ddc:000
    Language: English
    Type: reportzib , doc-type:preprint
    Format: application/pdf
    Library Location Call Number Volume/Issue/Year Availability
    BibTip Others were also interested in ...
  • 15
    Publication Date: 2020-11-16
    Description: Der Artikel gibt einen Einblick in das reiche Feld der Zusammenarbeit zwischen Mathematik und Medizin. Beispielhaft werden drei Erfolgsmodelle dargestellt: Medizinische Bildgebung, mathematische Modellierung und Biosignalverarbeitung im Bereich der Dynamik des Herzens sowie mathematische Modellierung und Simulation in der Krebstherapie Hyperthermie und der Mund-Kiefer-Gesichts-Chirurgie. In allen Fällen existiert ein Gleichklang der Interessen von Medizin und Mathematik: Beide Disziplinen wollen die Resultate schnell und zuverlässig. Für die Klinik heißt das, dass notwendige Rechnungen in möglichst kurzer Zeit, und zwar auf dem PC, ablaufen müssen und dass die Resultate so genau und belastbar sein müssen, dass medizinische Entscheidungen darauf aufbauen können. Für die Mathematik folgt daraus, dass höchste Anforderungen an die Effizienz der verwendeten Algorithmen und die darauf aufbauende Software in Numerik und Visualisierung zu stellen sind. Jedes Kapitel endet mit einer Darstellung der Perspektive des jeweiligen Gebietes. Abschließend werden mögliche Handlungsoptionen für Politik und Wirtschaft diskutiert.
    Keywords: ddc:000
    Language: German
    Type: reportzib , doc-type:preprint
    Format: application/pdf
    Library Location Call Number Volume/Issue/Year Availability
    BibTip Others were also interested in ...
  • 16
    Publication Date: 2022-03-14
    Description: This article introduces constraint integer programming (CIP), which is a novel way to combine constraint programming (CP) and mixed integer programming (MIP) methodologies. CIP is a generalization of MIP that supports the notion of general constraints as in CP. This approach is supported by the CIP framework SCIP, which also integrates techniques for solving satisfiability problems. SCIP is available in source code and free for noncommercial use. We demonstrate the usefulness of CIP on three tasks. First, we apply the constraint integer programming approach to pure mixed integer programs. Computational experiments show that SCIP is almost competitive to current state-of-the-art commercial MIP solvers. Second, we demonstrate how to use CIP techniques to compute the number of optimal solutions of integer programs. Third, we employ the CIP framework to solve chip design verification problems, which involve some highly nonlinear constraint types that are very hard to handle by pure MIP solvers. The CIP approach is very effective here: it can apply the full sophisticated MIP machinery to the linear part of the problem, while dealing with the nonlinear constraints by employing constraint programming techniques.
    Keywords: ddc:000
    Language: English
    Type: reportzib , doc-type:preprint
    Format: application/pdf
    Format: application/postscript
    Library Location Call Number Volume/Issue/Year Availability
    BibTip Others were also interested in ...
  • 17
    Publication Date: 2022-04-11
    Description: Mit dem Informationsdienst Math&Industry soll ein Prototyp eines dezentralen Informationssystems für Förderprogramme und Forschungsprojekte geschaffen werden, das sich auf andere Programme (bezüglich angewandter Mathematik, aber auch darüber hinaus) übertragen lässt. Das betrifft sowohl die Konzeption (Strukturierung der Informationen) als auch die Werkzeuge, die anderen BMBF-Förderprogrammen zur Verfügung gestellt werden können. Damit sollen diese in die Lage versetzt werden, die in Math&Industry entwickelten Konzepte und Werkzeuge an ihre spezifischen Bedürfnisse anzupassen.
    Keywords: ddc:000
    Language: German
    Type: reportzib , doc-type:preprint
    Format: application/pdf
    Library Location Call Number Volume/Issue/Year Availability
    BibTip Others were also interested in ...
Close ⊗
This website uses cookies and the analysis tool Matomo. More information can be found here...